var/home/core/zuul-output/0000755000175000017500000000000015113440661014526 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015113455362015476 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005203414715113455353017710 0ustar rootrootDec 02 01:36:22 crc systemd[1]: Starting Kubernetes Kubelet... Dec 02 01:36:22 crc restorecon[4692]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:22 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 01:36:23 crc restorecon[4692]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 02 01:36:23 crc restorecon[4692]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 02 01:36:23 crc kubenswrapper[4884]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 02 01:36:23 crc kubenswrapper[4884]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 02 01:36:23 crc kubenswrapper[4884]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 02 01:36:23 crc kubenswrapper[4884]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 02 01:36:23 crc kubenswrapper[4884]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 02 01:36:23 crc kubenswrapper[4884]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.441870 4884 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445101 4884 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445123 4884 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445129 4884 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445135 4884 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445140 4884 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445148 4884 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445156 4884 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445162 4884 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445168 4884 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445173 4884 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445179 4884 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445185 4884 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445190 4884 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445195 4884 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445200 4884 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445205 4884 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445210 4884 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445215 4884 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445220 4884 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445225 4884 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445229 4884 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445241 4884 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445246 4884 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445251 4884 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445255 4884 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445260 4884 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445266 4884 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445271 4884 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445276 4884 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445281 4884 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445287 4884 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445293 4884 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445298 4884 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445304 4884 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445309 4884 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445313 4884 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445318 4884 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445324 4884 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445329 4884 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445335 4884 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445340 4884 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445345 4884 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445349 4884 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445354 4884 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445359 4884 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445363 4884 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445368 4884 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445373 4884 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445390 4884 feature_gate.go:330] unrecognized feature gate: Example Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445396 4884 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445402 4884 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445407 4884 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445412 4884 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445417 4884 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445422 4884 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445427 4884 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445433 4884 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445440 4884 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445446 4884 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445451 4884 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445456 4884 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445461 4884 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445466 4884 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445471 4884 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445478 4884 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445483 4884 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445488 4884 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445493 4884 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445501 4884 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445506 4884 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.445511 4884 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445601 4884 flags.go:64] FLAG: --address="0.0.0.0" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445612 4884 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445621 4884 flags.go:64] FLAG: --anonymous-auth="true" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445629 4884 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445635 4884 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445641 4884 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445649 4884 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445657 4884 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445662 4884 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445668 4884 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445674 4884 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445680 4884 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445685 4884 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445691 4884 flags.go:64] FLAG: --cgroup-root="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445697 4884 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445703 4884 flags.go:64] FLAG: --client-ca-file="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445708 4884 flags.go:64] FLAG: --cloud-config="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445714 4884 flags.go:64] FLAG: --cloud-provider="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445720 4884 flags.go:64] FLAG: --cluster-dns="[]" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445766 4884 flags.go:64] FLAG: --cluster-domain="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445772 4884 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445778 4884 flags.go:64] FLAG: --config-dir="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445784 4884 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445790 4884 flags.go:64] FLAG: --container-log-max-files="5" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445798 4884 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445804 4884 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445809 4884 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445815 4884 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445821 4884 flags.go:64] FLAG: --contention-profiling="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445827 4884 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445835 4884 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445841 4884 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445847 4884 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445854 4884 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445860 4884 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445866 4884 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445872 4884 flags.go:64] FLAG: --enable-load-reader="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445878 4884 flags.go:64] FLAG: --enable-server="true" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445883 4884 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445892 4884 flags.go:64] FLAG: --event-burst="100" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445898 4884 flags.go:64] FLAG: --event-qps="50" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445903 4884 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445909 4884 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445914 4884 flags.go:64] FLAG: --eviction-hard="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445922 4884 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445928 4884 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445933 4884 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445939 4884 flags.go:64] FLAG: --eviction-soft="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445945 4884 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445950 4884 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445956 4884 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445962 4884 flags.go:64] FLAG: --experimental-mounter-path="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445967 4884 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445974 4884 flags.go:64] FLAG: --fail-swap-on="true" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445980 4884 flags.go:64] FLAG: --feature-gates="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445987 4884 flags.go:64] FLAG: --file-check-frequency="20s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445993 4884 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.445999 4884 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446006 4884 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446012 4884 flags.go:64] FLAG: --healthz-port="10248" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446018 4884 flags.go:64] FLAG: --help="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446023 4884 flags.go:64] FLAG: --hostname-override="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446029 4884 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446035 4884 flags.go:64] FLAG: --http-check-frequency="20s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446041 4884 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446048 4884 flags.go:64] FLAG: --image-credential-provider-config="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446053 4884 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446059 4884 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446064 4884 flags.go:64] FLAG: --image-service-endpoint="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446070 4884 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446075 4884 flags.go:64] FLAG: --kube-api-burst="100" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446081 4884 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446088 4884 flags.go:64] FLAG: --kube-api-qps="50" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446094 4884 flags.go:64] FLAG: --kube-reserved="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446099 4884 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446105 4884 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446111 4884 flags.go:64] FLAG: --kubelet-cgroups="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446116 4884 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446122 4884 flags.go:64] FLAG: --lock-file="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446127 4884 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446133 4884 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446139 4884 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446148 4884 flags.go:64] FLAG: --log-json-split-stream="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446153 4884 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446159 4884 flags.go:64] FLAG: --log-text-split-stream="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446164 4884 flags.go:64] FLAG: --logging-format="text" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446170 4884 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446176 4884 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446182 4884 flags.go:64] FLAG: --manifest-url="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446187 4884 flags.go:64] FLAG: --manifest-url-header="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446195 4884 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446201 4884 flags.go:64] FLAG: --max-open-files="1000000" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446209 4884 flags.go:64] FLAG: --max-pods="110" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446215 4884 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446222 4884 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446228 4884 flags.go:64] FLAG: --memory-manager-policy="None" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446234 4884 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446239 4884 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446245 4884 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446251 4884 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446263 4884 flags.go:64] FLAG: --node-status-max-images="50" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446269 4884 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446275 4884 flags.go:64] FLAG: --oom-score-adj="-999" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446281 4884 flags.go:64] FLAG: --pod-cidr="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446287 4884 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446296 4884 flags.go:64] FLAG: --pod-manifest-path="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446302 4884 flags.go:64] FLAG: --pod-max-pids="-1" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446308 4884 flags.go:64] FLAG: --pods-per-core="0" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446313 4884 flags.go:64] FLAG: --port="10250" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446319 4884 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446325 4884 flags.go:64] FLAG: --provider-id="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446331 4884 flags.go:64] FLAG: --qos-reserved="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446337 4884 flags.go:64] FLAG: --read-only-port="10255" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446342 4884 flags.go:64] FLAG: --register-node="true" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446348 4884 flags.go:64] FLAG: --register-schedulable="true" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446354 4884 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446364 4884 flags.go:64] FLAG: --registry-burst="10" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446370 4884 flags.go:64] FLAG: --registry-qps="5" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446375 4884 flags.go:64] FLAG: --reserved-cpus="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446381 4884 flags.go:64] FLAG: --reserved-memory="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446388 4884 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446394 4884 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446411 4884 flags.go:64] FLAG: --rotate-certificates="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446417 4884 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446422 4884 flags.go:64] FLAG: --runonce="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446428 4884 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446434 4884 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446440 4884 flags.go:64] FLAG: --seccomp-default="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446446 4884 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446452 4884 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446459 4884 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446464 4884 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446470 4884 flags.go:64] FLAG: --storage-driver-password="root" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446476 4884 flags.go:64] FLAG: --storage-driver-secure="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446482 4884 flags.go:64] FLAG: --storage-driver-table="stats" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446487 4884 flags.go:64] FLAG: --storage-driver-user="root" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446493 4884 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446499 4884 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446506 4884 flags.go:64] FLAG: --system-cgroups="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446512 4884 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446521 4884 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446527 4884 flags.go:64] FLAG: --tls-cert-file="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446532 4884 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446540 4884 flags.go:64] FLAG: --tls-min-version="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446546 4884 flags.go:64] FLAG: --tls-private-key-file="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446551 4884 flags.go:64] FLAG: --topology-manager-policy="none" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446556 4884 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446562 4884 flags.go:64] FLAG: --topology-manager-scope="container" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446568 4884 flags.go:64] FLAG: --v="2" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446575 4884 flags.go:64] FLAG: --version="false" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446582 4884 flags.go:64] FLAG: --vmodule="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446589 4884 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.446595 4884 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446754 4884 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446763 4884 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446770 4884 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446776 4884 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446782 4884 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446788 4884 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446797 4884 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446802 4884 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446808 4884 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446812 4884 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446817 4884 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446824 4884 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446830 4884 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446836 4884 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446841 4884 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446846 4884 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446851 4884 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446856 4884 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446861 4884 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446866 4884 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446871 4884 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446875 4884 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446880 4884 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446885 4884 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446890 4884 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446895 4884 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446900 4884 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446904 4884 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446909 4884 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446914 4884 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446918 4884 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446924 4884 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446928 4884 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446935 4884 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446940 4884 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446944 4884 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446949 4884 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446954 4884 feature_gate.go:330] unrecognized feature gate: Example Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446960 4884 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446965 4884 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446970 4884 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446974 4884 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446979 4884 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446984 4884 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446989 4884 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446994 4884 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.446999 4884 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447005 4884 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447016 4884 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447022 4884 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447027 4884 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447031 4884 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447037 4884 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447041 4884 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447046 4884 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447051 4884 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447055 4884 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447060 4884 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447065 4884 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447070 4884 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447076 4884 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447081 4884 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447086 4884 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447092 4884 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447099 4884 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447105 4884 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447110 4884 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447115 4884 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447120 4884 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447126 4884 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.447134 4884 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.447293 4884 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.459783 4884 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.459839 4884 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460246 4884 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460265 4884 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460271 4884 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460276 4884 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460284 4884 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460291 4884 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460297 4884 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460303 4884 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460309 4884 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460315 4884 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460320 4884 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460325 4884 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460330 4884 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460335 4884 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460340 4884 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460345 4884 feature_gate.go:330] unrecognized feature gate: Example Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460350 4884 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460355 4884 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460360 4884 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460365 4884 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460370 4884 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460375 4884 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460381 4884 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460387 4884 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460392 4884 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460397 4884 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460403 4884 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460411 4884 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460418 4884 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460423 4884 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460429 4884 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460433 4884 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460438 4884 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460444 4884 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460449 4884 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460454 4884 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460458 4884 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460463 4884 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460469 4884 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460476 4884 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460482 4884 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460490 4884 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460496 4884 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460501 4884 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460507 4884 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460513 4884 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460518 4884 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460523 4884 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460530 4884 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460536 4884 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460541 4884 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460547 4884 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460552 4884 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460557 4884 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460562 4884 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460567 4884 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460572 4884 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460577 4884 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460582 4884 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460586 4884 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460592 4884 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460596 4884 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460601 4884 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460606 4884 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460610 4884 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460616 4884 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460620 4884 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460625 4884 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460630 4884 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460635 4884 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460640 4884 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.460648 4884 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460829 4884 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460841 4884 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460848 4884 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460853 4884 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460859 4884 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460864 4884 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460869 4884 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460874 4884 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460879 4884 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460885 4884 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460890 4884 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460896 4884 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460903 4884 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460908 4884 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460914 4884 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460919 4884 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460924 4884 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460930 4884 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460935 4884 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460957 4884 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460963 4884 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460968 4884 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460973 4884 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460978 4884 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460984 4884 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460989 4884 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460993 4884 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.460998 4884 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461003 4884 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461008 4884 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461014 4884 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461019 4884 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461024 4884 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461029 4884 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461035 4884 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461039 4884 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461044 4884 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461051 4884 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461056 4884 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461062 4884 feature_gate.go:330] unrecognized feature gate: Example Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461067 4884 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461071 4884 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461076 4884 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461081 4884 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461087 4884 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461092 4884 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461097 4884 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461102 4884 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461107 4884 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461111 4884 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461117 4884 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461121 4884 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461127 4884 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461132 4884 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461139 4884 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461146 4884 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461153 4884 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461159 4884 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461165 4884 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461171 4884 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461178 4884 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461184 4884 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461190 4884 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461195 4884 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461200 4884 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461206 4884 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461212 4884 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461217 4884 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461222 4884 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461227 4884 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.461232 4884 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.461239 4884 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.461717 4884 server.go:940] "Client rotation is on, will bootstrap in background" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.465308 4884 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.465396 4884 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.465985 4884 server.go:997] "Starting client certificate rotation" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.466011 4884 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.466395 4884 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-13 20:16:00.676290013 +0000 UTC Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.466534 4884 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1026h39m37.209774427s for next certificate rotation Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.471249 4884 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.472987 4884 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.480831 4884 log.go:25] "Validated CRI v1 runtime API" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.505124 4884 log.go:25] "Validated CRI v1 image API" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.507322 4884 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.510512 4884 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-02-01-31-08-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.510541 4884 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:41 fsType:tmpfs blockSize:0}] Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.529875 4884 manager.go:217] Machine: {Timestamp:2025-12-02 01:36:23.528429791 +0000 UTC m=+0.204266685 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:09169ac4-b9cd-4f4b-a571-21877303afe5 BootID:54af654b-c7be-43b5-9dfc-8a2692aad92d Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:41 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:cf:bf:5b Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:cf:bf:5b Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:8c:f9:ca Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:5d:b9:50 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:4d:95:28 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:4e:cf:3f Speed:-1 Mtu:1496} {Name:eth10 MacAddress:2a:f2:04:16:f8:c2 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:12:20:09:d4:6c:1d Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.530152 4884 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.530304 4884 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.530832 4884 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.531096 4884 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.531140 4884 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.531505 4884 topology_manager.go:138] "Creating topology manager with none policy" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.531533 4884 container_manager_linux.go:303] "Creating device plugin manager" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.531725 4884 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.531791 4884 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.532000 4884 state_mem.go:36] "Initialized new in-memory state store" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.532094 4884 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.533056 4884 kubelet.go:418] "Attempting to sync node with API server" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.533082 4884 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.533118 4884 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.533135 4884 kubelet.go:324] "Adding apiserver pod source" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.533149 4884 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.540151 4884 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.540276 4884 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Dec 02 01:36:23 crc kubenswrapper[4884]: E1202 01:36:23.540389 4884 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.540507 4884 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Dec 02 01:36:23 crc kubenswrapper[4884]: E1202 01:36:23.540620 4884 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.540670 4884 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.541840 4884 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.542803 4884 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.542849 4884 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.542865 4884 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.542880 4884 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.542902 4884 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.542916 4884 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.542930 4884 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.542952 4884 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.542969 4884 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.542982 4884 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.543008 4884 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.543022 4884 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.543523 4884 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.544217 4884 server.go:1280] "Started kubelet" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.544615 4884 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.544715 4884 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.544925 4884 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Dec 02 01:36:23 crc systemd[1]: Started Kubernetes Kubelet. Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.547462 4884 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.547530 4884 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.547616 4884 server.go:460] "Adding debug handlers to kubelet server" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.548193 4884 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 00:04:34.31607836 +0000 UTC Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.548273 4884 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1126h28m10.767810403s for next certificate rotation Dec 02 01:36:23 crc kubenswrapper[4884]: E1202 01:36:23.548903 4884 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.32:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187d422b7cd09df5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-02 01:36:23.544176117 +0000 UTC m=+0.220013041,LastTimestamp:2025-12-02 01:36:23.544176117 +0000 UTC m=+0.220013041,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.549371 4884 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.549408 4884 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.549384 4884 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.549949 4884 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Dec 02 01:36:23 crc kubenswrapper[4884]: E1202 01:36:23.550081 4884 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 02 01:36:23 crc kubenswrapper[4884]: E1202 01:36:23.550080 4884 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.550333 4884 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.550736 4884 factory.go:55] Registering systemd factory Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.550800 4884 factory.go:221] Registration of the systemd container factory successfully Dec 02 01:36:23 crc kubenswrapper[4884]: E1202 01:36:23.550720 4884 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="200ms" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.551423 4884 factory.go:153] Registering CRI-O factory Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.551640 4884 factory.go:221] Registration of the crio container factory successfully Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.551898 4884 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.551947 4884 factory.go:103] Registering Raw factory Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.552063 4884 manager.go:1196] Started watching for new ooms in manager Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.556056 4884 manager.go:319] Starting recovery of all containers Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.568724 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.568854 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.568879 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.568903 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.568925 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.568944 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.568963 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.568982 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569006 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569026 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569047 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569066 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569093 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569119 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569138 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569157 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569178 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569196 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569214 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569233 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569252 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569272 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569292 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569312 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569332 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569352 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569375 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569395 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569415 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569434 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569453 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569483 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569502 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569524 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569545 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569563 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569581 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569601 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569623 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569642 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569661 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569679 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569698 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569718 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569737 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569795 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569816 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569839 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569859 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569879 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569898 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569922 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.569988 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570013 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570033 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570054 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570077 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570100 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570123 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570143 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570166 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570188 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570207 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570227 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570246 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570267 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570289 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570315 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570334 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570355 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570377 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570396 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570415 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570434 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570455 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570475 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570497 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570516 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570536 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570555 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570573 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570592 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570614 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570635 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570655 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570675 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570698 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570723 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570800 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570845 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570875 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570903 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570975 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.570998 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571017 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571038 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571057 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571083 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571104 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571125 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571145 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571166 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571185 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571209 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571242 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571263 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571284 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571306 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571328 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571349 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571371 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571394 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571415 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571440 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571461 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571482 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571502 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571523 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571542 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571561 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571583 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571605 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571626 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571644 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571663 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571683 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571702 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571721 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571832 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571865 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571889 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571910 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571928 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571946 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571964 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.571985 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572007 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572028 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572047 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572067 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572085 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572107 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572126 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572148 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572169 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572193 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572213 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572234 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572252 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572270 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572290 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572311 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572330 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572349 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572370 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572391 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572417 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572444 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572471 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572497 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572517 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.572541 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.573587 4884 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.573646 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.573680 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.573718 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.573782 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.573818 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.573873 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.573911 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.573941 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.573968 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.573992 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574073 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574107 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574135 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574164 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574201 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574229 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574256 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574284 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574312 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574338 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574366 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574393 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574419 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574444 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574468 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574492 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574516 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574540 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574567 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574591 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574618 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574645 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574669 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574695 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574728 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574786 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574816 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574842 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574871 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574895 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574920 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574943 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574966 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.574992 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.575014 4884 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.575035 4884 reconstruct.go:97] "Volume reconstruction finished" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.575052 4884 reconciler.go:26] "Reconciler: start to sync state" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.603359 4884 manager.go:324] Recovery completed Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.609920 4884 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.612797 4884 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.612847 4884 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.612884 4884 kubelet.go:2335] "Starting kubelet main sync loop" Dec 02 01:36:23 crc kubenswrapper[4884]: E1202 01:36:23.613021 4884 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 02 01:36:23 crc kubenswrapper[4884]: W1202 01:36:23.615051 4884 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Dec 02 01:36:23 crc kubenswrapper[4884]: E1202 01:36:23.615159 4884 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.618791 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.620440 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.620491 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.620505 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.621568 4884 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.621596 4884 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.621623 4884 state_mem.go:36] "Initialized new in-memory state store" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.632895 4884 policy_none.go:49] "None policy: Start" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.633670 4884 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.633707 4884 state_mem.go:35] "Initializing new in-memory state store" Dec 02 01:36:23 crc kubenswrapper[4884]: E1202 01:36:23.651150 4884 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.697003 4884 manager.go:334] "Starting Device Plugin manager" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.697607 4884 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.697658 4884 server.go:79] "Starting device plugin registration server" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.698386 4884 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.698408 4884 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.698866 4884 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.698989 4884 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.698999 4884 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.713229 4884 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.713307 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:23 crc kubenswrapper[4884]: E1202 01:36:23.713310 4884 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.714616 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.714653 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.714665 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.714804 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.714962 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.715037 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.715578 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.715625 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.715644 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.715846 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.715976 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.716018 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.716266 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.716300 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.716312 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.718308 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.718357 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.718373 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.718615 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.719457 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.719574 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.720387 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.720415 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.720433 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.720629 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.721089 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.721156 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.721220 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.721249 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.721260 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.723561 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.723611 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.723630 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.724021 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.724043 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.724055 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.724330 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.724381 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.724774 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.724828 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.724839 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.725303 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.725338 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.725351 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:23 crc kubenswrapper[4884]: E1202 01:36:23.751554 4884 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="400ms" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.777721 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.777822 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.777860 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.777948 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.778051 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.778097 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.778142 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.778200 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.778224 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.778641 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.779283 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.779344 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.779369 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.779415 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.779444 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.798781 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.800561 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.800626 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.800646 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.800694 4884 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 01:36:23 crc kubenswrapper[4884]: E1202 01:36:23.801529 4884 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.32:6443: connect: connection refused" node="crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881095 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881247 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881289 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881351 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881363 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881393 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881486 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881559 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881590 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881552 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881599 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881664 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881674 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881722 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881792 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881824 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881888 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881722 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881838 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881991 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881891 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881885 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.881931 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.882106 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.882161 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.882176 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.882181 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.882209 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.882247 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:23 crc kubenswrapper[4884]: I1202 01:36:23.882274 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.002507 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.004004 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.004044 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.004062 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.004097 4884 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 01:36:24 crc kubenswrapper[4884]: E1202 01:36:24.004641 4884 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.32:6443: connect: connection refused" node="crc" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.048124 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.054270 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.079637 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.086988 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.093246 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 01:36:24 crc kubenswrapper[4884]: W1202 01:36:24.099977 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-06240d344f0c304a300d8249d31ca582c848ceb31e9f9d1c7b816ac8b2a758b9 WatchSource:0}: Error finding container 06240d344f0c304a300d8249d31ca582c848ceb31e9f9d1c7b816ac8b2a758b9: Status 404 returned error can't find the container with id 06240d344f0c304a300d8249d31ca582c848ceb31e9f9d1c7b816ac8b2a758b9 Dec 02 01:36:24 crc kubenswrapper[4884]: W1202 01:36:24.101157 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-b0994152f71592fd7aee01123caf3e2377196b16f84b590fabb1249206384ec6 WatchSource:0}: Error finding container b0994152f71592fd7aee01123caf3e2377196b16f84b590fabb1249206384ec6: Status 404 returned error can't find the container with id b0994152f71592fd7aee01123caf3e2377196b16f84b590fabb1249206384ec6 Dec 02 01:36:24 crc kubenswrapper[4884]: W1202 01:36:24.107860 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-3e221b561b4067890df60ede6f32779c995e033f293af5b4fc16633fcc7f7def WatchSource:0}: Error finding container 3e221b561b4067890df60ede6f32779c995e033f293af5b4fc16633fcc7f7def: Status 404 returned error can't find the container with id 3e221b561b4067890df60ede6f32779c995e033f293af5b4fc16633fcc7f7def Dec 02 01:36:24 crc kubenswrapper[4884]: W1202 01:36:24.123106 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-e14dc203f8b5ab0664a6e0afefa1cd76167fb35c1fc55ac4a91e3cbc27924aad WatchSource:0}: Error finding container e14dc203f8b5ab0664a6e0afefa1cd76167fb35c1fc55ac4a91e3cbc27924aad: Status 404 returned error can't find the container with id e14dc203f8b5ab0664a6e0afefa1cd76167fb35c1fc55ac4a91e3cbc27924aad Dec 02 01:36:24 crc kubenswrapper[4884]: W1202 01:36:24.128899 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-1690b3bd590abc2bf795b5bfc7ce89848fa45662b3db0b12dce7c5b139b1ffcf WatchSource:0}: Error finding container 1690b3bd590abc2bf795b5bfc7ce89848fa45662b3db0b12dce7c5b139b1ffcf: Status 404 returned error can't find the container with id 1690b3bd590abc2bf795b5bfc7ce89848fa45662b3db0b12dce7c5b139b1ffcf Dec 02 01:36:24 crc kubenswrapper[4884]: E1202 01:36:24.152884 4884 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="800ms" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.405800 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.408369 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.408414 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.408427 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.408472 4884 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 01:36:24 crc kubenswrapper[4884]: E1202 01:36:24.409030 4884 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.32:6443: connect: connection refused" node="crc" Dec 02 01:36:24 crc kubenswrapper[4884]: W1202 01:36:24.419487 4884 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Dec 02 01:36:24 crc kubenswrapper[4884]: E1202 01:36:24.419596 4884 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Dec 02 01:36:24 crc kubenswrapper[4884]: W1202 01:36:24.457860 4884 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Dec 02 01:36:24 crc kubenswrapper[4884]: E1202 01:36:24.457980 4884 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.546414 4884 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Dec 02 01:36:24 crc kubenswrapper[4884]: W1202 01:36:24.572842 4884 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Dec 02 01:36:24 crc kubenswrapper[4884]: E1202 01:36:24.572942 4884 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.621682 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353"} Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.621897 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1690b3bd590abc2bf795b5bfc7ce89848fa45662b3db0b12dce7c5b139b1ffcf"} Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.622011 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.623543 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.623601 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.623622 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.623762 4884 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7" exitCode=0 Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.623851 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7"} Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.623900 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e14dc203f8b5ab0664a6e0afefa1cd76167fb35c1fc55ac4a91e3cbc27924aad"} Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.624040 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.624738 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.624807 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.624825 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.625519 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7"} Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.625546 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3e221b561b4067890df60ede6f32779c995e033f293af5b4fc16633fcc7f7def"} Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.627201 4884 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7" exitCode=0 Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.627238 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7"} Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.627279 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"06240d344f0c304a300d8249d31ca582c848ceb31e9f9d1c7b816ac8b2a758b9"} Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.627399 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.628409 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.628455 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.628469 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.629909 4884 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="2d85d688f8e8f22166407638c3e8e8c2232dddd46b4a601e3ebb4defc19e134d" exitCode=0 Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.629957 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"2d85d688f8e8f22166407638c3e8e8c2232dddd46b4a601e3ebb4defc19e134d"} Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.629988 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"b0994152f71592fd7aee01123caf3e2377196b16f84b590fabb1249206384ec6"} Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.630079 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.630993 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.631021 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.631032 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.634955 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.637490 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.637534 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:24 crc kubenswrapper[4884]: I1202 01:36:24.637545 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:24 crc kubenswrapper[4884]: E1202 01:36:24.953721 4884 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="1.6s" Dec 02 01:36:25 crc kubenswrapper[4884]: W1202 01:36:25.087297 4884 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.32:6443: connect: connection refused Dec 02 01:36:25 crc kubenswrapper[4884]: E1202 01:36:25.087400 4884 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.32:6443: connect: connection refused" logger="UnhandledError" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.209111 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.210134 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.210168 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.210177 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.210206 4884 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 01:36:25 crc kubenswrapper[4884]: E1202 01:36:25.210476 4884 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.32:6443: connect: connection refused" node="crc" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.634888 4884 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353" exitCode=0 Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.634940 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353"} Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.634965 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"67a218e1965e88c816e5dc27a6bfe7c6043522794f68c4eca80982082c1c3b62"} Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.634976 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"21ab31a5e5a2b1ae8b6b2d1ee1dcc1aefcc6efe942947fa71b801f90dc346f1f"} Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.634985 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"27f06f9f8e8f076b46dd2ec9d609217fa18bc94c8f2cc7bf6a6e530820f74e92"} Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.635068 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.636311 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.636334 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.636344 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.637811 4884 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36" exitCode=0 Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.637847 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36"} Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.637918 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.638505 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.638524 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.638533 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.653759 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f"} Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.653821 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0"} Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.653835 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a"} Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.654004 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.655240 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.655266 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.655278 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.659658 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310"} Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.659718 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8"} Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.659736 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234"} Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.661923 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"5ed1e538b1193a685d075ad12cba91a17fe0f3e07fe1d5678f687b0385d319bc"} Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.662039 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.662938 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.662975 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:25 crc kubenswrapper[4884]: I1202 01:36:25.662985 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.670173 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db"} Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.670234 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.670268 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf"} Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.671211 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.671251 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.671260 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.673242 4884 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f" exitCode=0 Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.673313 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.673403 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.673391 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f"} Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.674065 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.674090 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.674100 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.674384 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.674430 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.674444 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.810844 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.812831 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.812892 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.812915 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:26 crc kubenswrapper[4884]: I1202 01:36:26.812959 4884 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 01:36:27 crc kubenswrapper[4884]: I1202 01:36:27.285226 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:27 crc kubenswrapper[4884]: I1202 01:36:27.680447 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f"} Dec 02 01:36:27 crc kubenswrapper[4884]: I1202 01:36:27.680509 4884 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 01:36:27 crc kubenswrapper[4884]: I1202 01:36:27.680518 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3"} Dec 02 01:36:27 crc kubenswrapper[4884]: I1202 01:36:27.680540 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4"} Dec 02 01:36:27 crc kubenswrapper[4884]: I1202 01:36:27.680553 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:27 crc kubenswrapper[4884]: I1202 01:36:27.681418 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:27 crc kubenswrapper[4884]: I1202 01:36:27.681490 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:27 crc kubenswrapper[4884]: I1202 01:36:27.681510 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.289391 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.289587 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.291318 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.291375 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.291393 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.692845 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7"} Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.692899 4884 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.693004 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.692910 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77"} Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.694469 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.698965 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.699053 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.699080 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.699857 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.700037 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:28 crc kubenswrapper[4884]: I1202 01:36:28.700169 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:29 crc kubenswrapper[4884]: I1202 01:36:29.696096 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:29 crc kubenswrapper[4884]: I1202 01:36:29.697471 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:29 crc kubenswrapper[4884]: I1202 01:36:29.697529 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:29 crc kubenswrapper[4884]: I1202 01:36:29.697547 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:29 crc kubenswrapper[4884]: I1202 01:36:29.992503 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.067378 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.067608 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.069273 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.069313 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.069325 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.148196 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.148804 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.150498 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.150558 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.150580 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.618722 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.619309 4884 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.619505 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.621181 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.621234 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.621252 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.699457 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.701386 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.701453 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:30 crc kubenswrapper[4884]: I1202 01:36:30.701472 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:31 crc kubenswrapper[4884]: I1202 01:36:31.475878 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:31 crc kubenswrapper[4884]: I1202 01:36:31.476148 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:31 crc kubenswrapper[4884]: I1202 01:36:31.477880 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:31 crc kubenswrapper[4884]: I1202 01:36:31.477954 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:31 crc kubenswrapper[4884]: I1202 01:36:31.477975 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:32 crc kubenswrapper[4884]: I1202 01:36:32.486333 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:32 crc kubenswrapper[4884]: I1202 01:36:32.487242 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:32 crc kubenswrapper[4884]: I1202 01:36:32.488964 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:32 crc kubenswrapper[4884]: I1202 01:36:32.489020 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:32 crc kubenswrapper[4884]: I1202 01:36:32.489037 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:33 crc kubenswrapper[4884]: I1202 01:36:33.237919 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 02 01:36:33 crc kubenswrapper[4884]: I1202 01:36:33.238252 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:33 crc kubenswrapper[4884]: I1202 01:36:33.240824 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:33 crc kubenswrapper[4884]: I1202 01:36:33.240904 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:33 crc kubenswrapper[4884]: I1202 01:36:33.240923 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:33 crc kubenswrapper[4884]: E1202 01:36:33.713509 4884 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 02 01:36:34 crc kubenswrapper[4884]: I1202 01:36:34.476162 4884 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 02 01:36:34 crc kubenswrapper[4884]: I1202 01:36:34.476296 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 01:36:34 crc kubenswrapper[4884]: I1202 01:36:34.994876 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:34 crc kubenswrapper[4884]: I1202 01:36:34.995050 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:34 crc kubenswrapper[4884]: I1202 01:36:34.996309 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:34 crc kubenswrapper[4884]: I1202 01:36:34.996339 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:34 crc kubenswrapper[4884]: I1202 01:36:34.996347 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:35 crc kubenswrapper[4884]: I1202 01:36:35.004252 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:35 crc kubenswrapper[4884]: I1202 01:36:35.546513 4884 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 02 01:36:35 crc kubenswrapper[4884]: I1202 01:36:35.718798 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:35 crc kubenswrapper[4884]: I1202 01:36:35.720102 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:35 crc kubenswrapper[4884]: I1202 01:36:35.720131 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:35 crc kubenswrapper[4884]: I1202 01:36:35.720139 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:35 crc kubenswrapper[4884]: I1202 01:36:35.723926 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:36 crc kubenswrapper[4884]: W1202 01:36:36.366600 4884 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 02 01:36:36 crc kubenswrapper[4884]: I1202 01:36:36.366730 4884 trace.go:236] Trace[368473276]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Dec-2025 01:36:26.364) (total time: 10001ms): Dec 02 01:36:36 crc kubenswrapper[4884]: Trace[368473276]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (01:36:36.366) Dec 02 01:36:36 crc kubenswrapper[4884]: Trace[368473276]: [10.001823291s] [10.001823291s] END Dec 02 01:36:36 crc kubenswrapper[4884]: E1202 01:36:36.366784 4884 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 02 01:36:36 crc kubenswrapper[4884]: E1202 01:36:36.555084 4884 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 02 01:36:36 crc kubenswrapper[4884]: I1202 01:36:36.722437 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:36 crc kubenswrapper[4884]: I1202 01:36:36.724224 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:36 crc kubenswrapper[4884]: I1202 01:36:36.724288 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:36 crc kubenswrapper[4884]: I1202 01:36:36.724332 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:36 crc kubenswrapper[4884]: E1202 01:36:36.814467 4884 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 02 01:36:36 crc kubenswrapper[4884]: W1202 01:36:36.859110 4884 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 02 01:36:36 crc kubenswrapper[4884]: I1202 01:36:36.859254 4884 trace.go:236] Trace[1740261232]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Dec-2025 01:36:26.857) (total time: 10001ms): Dec 02 01:36:36 crc kubenswrapper[4884]: Trace[1740261232]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (01:36:36.859) Dec 02 01:36:36 crc kubenswrapper[4884]: Trace[1740261232]: [10.00175879s] [10.00175879s] END Dec 02 01:36:36 crc kubenswrapper[4884]: E1202 01:36:36.859290 4884 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 02 01:36:36 crc kubenswrapper[4884]: I1202 01:36:36.981625 4884 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 02 01:36:36 crc kubenswrapper[4884]: I1202 01:36:36.981719 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 02 01:36:36 crc kubenswrapper[4884]: I1202 01:36:36.994905 4884 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 02 01:36:36 crc kubenswrapper[4884]: I1202 01:36:36.994979 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 02 01:36:37 crc kubenswrapper[4884]: I1202 01:36:37.296083 4884 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]log ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]etcd ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/generic-apiserver-start-informers ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/priority-and-fairness-filter ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/start-apiextensions-informers ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/start-apiextensions-controllers ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/crd-informer-synced ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/start-system-namespaces-controller ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 02 01:36:37 crc kubenswrapper[4884]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 02 01:36:37 crc kubenswrapper[4884]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/bootstrap-controller ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/start-kube-aggregator-informers ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/apiservice-registration-controller ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/apiservice-discovery-controller ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]autoregister-completion ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/apiservice-openapi-controller ok Dec 02 01:36:37 crc kubenswrapper[4884]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 02 01:36:37 crc kubenswrapper[4884]: livez check failed Dec 02 01:36:37 crc kubenswrapper[4884]: I1202 01:36:37.296177 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:36:40 crc kubenswrapper[4884]: I1202 01:36:40.015142 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:40 crc kubenswrapper[4884]: I1202 01:36:40.018121 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:40 crc kubenswrapper[4884]: I1202 01:36:40.018270 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:40 crc kubenswrapper[4884]: I1202 01:36:40.018376 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:40 crc kubenswrapper[4884]: I1202 01:36:40.018497 4884 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 01:36:40 crc kubenswrapper[4884]: E1202 01:36:40.023271 4884 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 02 01:36:41 crc kubenswrapper[4884]: I1202 01:36:41.959076 4884 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 02 01:36:41 crc kubenswrapper[4884]: I1202 01:36:41.988704 4884 trace.go:236] Trace[639857042]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Dec-2025 01:36:27.091) (total time: 14896ms): Dec 02 01:36:41 crc kubenswrapper[4884]: Trace[639857042]: ---"Objects listed" error: 14896ms (01:36:41.988) Dec 02 01:36:41 crc kubenswrapper[4884]: Trace[639857042]: [14.89682115s] [14.89682115s] END Dec 02 01:36:41 crc kubenswrapper[4884]: I1202 01:36:41.988777 4884 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 02 01:36:41 crc kubenswrapper[4884]: I1202 01:36:41.990568 4884 trace.go:236] Trace[1431498269]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Dec-2025 01:36:27.714) (total time: 14275ms): Dec 02 01:36:41 crc kubenswrapper[4884]: Trace[1431498269]: ---"Objects listed" error: 14275ms (01:36:41.990) Dec 02 01:36:41 crc kubenswrapper[4884]: Trace[1431498269]: [14.275523049s] [14.275523049s] END Dec 02 01:36:41 crc kubenswrapper[4884]: I1202 01:36:41.990612 4884 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 02 01:36:41 crc kubenswrapper[4884]: I1202 01:36:41.995315 4884 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.030096 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.046129 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.292459 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.293215 4884 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.293272 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.297326 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.486870 4884 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.486939 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.547030 4884 apiserver.go:52] "Watching apiserver" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.550028 4884 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.550390 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.550894 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.550969 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.550895 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.551052 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.551056 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.551253 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.551140 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.551338 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.551262 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.555027 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.555349 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.555434 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.555432 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.555432 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.555700 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.556521 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.562891 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.563383 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.581983 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.594831 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.608208 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.622119 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.636993 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.650917 4884 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.657076 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-jc7vx"] Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.657632 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.659983 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.660424 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-p545z"] Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.660852 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.660880 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-p545z" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.661006 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.661101 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.662536 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.663000 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.663464 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.664975 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.665471 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.678436 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.690476 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.699608 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.699654 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.699682 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.699711 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.699735 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.699775 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.699801 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.699822 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.699847 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.699872 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.699899 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.699954 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.699967 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.699974 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700024 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700043 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700061 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700078 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700095 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700113 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700128 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700142 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700158 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700144 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700162 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700175 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700212 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700211 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700269 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700299 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700322 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700346 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700368 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700417 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700439 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700462 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700488 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700511 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700533 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700587 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700612 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700654 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700674 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700697 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700720 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700760 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700783 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700806 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700827 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700852 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700874 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700895 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700915 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700934 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700958 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700979 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701001 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701028 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701073 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701097 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701117 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701139 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701160 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701181 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701201 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701223 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701243 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701263 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701286 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701307 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701329 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701349 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701369 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701390 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701410 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701440 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701464 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701486 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701510 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700321 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700417 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700523 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700824 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.700927 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701026 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701033 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701105 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701271 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701342 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701635 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701389 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701414 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701551 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701719 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701836 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701886 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.702032 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.702063 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.702076 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.702278 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.702419 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.702484 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.702493 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.702677 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.702908 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704124 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.701556 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704348 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704397 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704424 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704449 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704470 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704509 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704536 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704557 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704576 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704597 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704618 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704636 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704654 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704675 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704694 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704712 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704735 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704780 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704803 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704821 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704839 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704859 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704880 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704907 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704927 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704950 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704971 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.704994 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705013 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705032 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705052 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705073 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705094 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705112 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705131 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705149 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705544 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705570 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705601 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705689 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705782 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705821 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705849 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705876 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705897 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705924 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705944 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705978 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706007 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706028 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706046 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706064 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706093 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706110 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706130 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706155 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706189 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706210 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706235 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706255 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706188 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706277 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706557 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706607 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706685 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706738 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706811 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706853 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706896 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.707440 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.707490 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.707525 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.707563 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.708555 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.708716 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.708793 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.708830 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.708867 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.708903 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.708942 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.708986 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709024 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709067 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709103 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709138 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709180 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709217 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709256 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709292 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709340 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709375 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709413 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709449 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709485 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709521 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709556 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709600 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709646 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709682 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709718 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709821 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709869 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709907 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.710557 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.710617 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.710655 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.710695 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.710731 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.710793 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.710826 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.710871 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.710907 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.710946 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.711006 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.711042 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.711082 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.711119 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.711156 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.711193 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.711230 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.711355 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.711441 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.711612 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.716730 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721430 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721500 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721539 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721572 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721601 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721629 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721657 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721686 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721713 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721757 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721781 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721808 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721905 4884 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721921 4884 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721935 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721950 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721966 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721977 4884 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721989 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722006 4884 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722018 4884 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722031 4884 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722046 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722063 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722075 4884 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722085 4884 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722096 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722111 4884 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722124 4884 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722136 4884 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722153 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722164 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722175 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722185 4884 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722198 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722209 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722218 4884 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722230 4884 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722243 4884 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722254 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722265 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722275 4884 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722288 4884 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722299 4884 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722312 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722594 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.723088 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.723934 4884 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705573 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.705798 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.726268 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706416 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706441 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.706493 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.708199 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.708266 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.708327 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.708318 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.708896 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.708918 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709201 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709301 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709377 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.709816 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.710190 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.710307 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.710395 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.711260 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.711534 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.711621 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.716967 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.716957 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.717159 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.717280 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.717402 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.717553 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.717913 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.717984 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.718247 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.718599 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.718630 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.718693 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.719019 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.719279 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.719321 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.719508 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.719514 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.719511 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.719589 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.719583 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.719763 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.719892 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.719920 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.720327 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.717427 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.720590 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721019 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721277 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.720507 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721623 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721777 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.721976 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722024 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722063 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722424 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722589 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722794 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722821 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722834 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.722861 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.725246 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.725410 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.725667 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.725783 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.725947 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.726029 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.727255 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.727701 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.727959 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.728407 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.728731 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.728790 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.728983 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:36:43.228955575 +0000 UTC m=+19.904792469 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.729048 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.729127 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.729161 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.729409 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.729423 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.729802 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.730063 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.730422 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.730429 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.730701 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.731110 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.731894 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.731892 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.731994 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.732013 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.732073 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.732644 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.732638 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.732973 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.733045 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.730718 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.733105 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.734210 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.734271 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.734736 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.733937 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.735056 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.733118 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.733399 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.733410 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.733455 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.731519 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.733572 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.733595 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.733643 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.733666 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.732346 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.735346 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.734527 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.735581 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.735923 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.736684 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.736703 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.737271 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.737416 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.737682 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.738278 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.747711 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.740157 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.740597 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.742946 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.740994 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.752671 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.752699 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.752976 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.753354 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.741551 4884 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.755073 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:43.25505477 +0000 UTC m=+19.930891654 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.739476 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.753878 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.754394 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.754230 4884 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.754612 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.755336 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.755348 4884 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.755377 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.755387 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:43.255379918 +0000 UTC m=+19.931216802 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.755633 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:43.255613924 +0000 UTC m=+19.931450828 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.756019 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.757390 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.757503 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.757849 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.757906 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.759066 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.760167 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.761004 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.761628 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.761792 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.762574 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.746559 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.745639 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.765364 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.765600 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.765667 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.765698 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.765726 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.766801 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.768487 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.768914 4884 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db" exitCode=255 Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.770980 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.772629 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.772769 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.773205 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.773233 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.773361 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.738375 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.775367 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.788292 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.788329 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.787869 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.797099 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.797135 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.797151 4884 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.797212 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:43.297190856 +0000 UTC m=+19.973027740 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.797294 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.800657 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db"} Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.803513 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.805272 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.821000 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826199 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/9385a5c9-a156-460a-8ca1-9b342ab835c9-rootfs\") pod \"machine-config-daemon-jc7vx\" (UID: \"9385a5c9-a156-460a-8ca1-9b342ab835c9\") " pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826242 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826261 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9385a5c9-a156-460a-8ca1-9b342ab835c9-proxy-tls\") pod \"machine-config-daemon-jc7vx\" (UID: \"9385a5c9-a156-460a-8ca1-9b342ab835c9\") " pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826277 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/531c7e95-83b1-4ad1-ad0e-490d9520e8d6-hosts-file\") pod \"node-resolver-p545z\" (UID: \"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\") " pod="openshift-dns/node-resolver-p545z" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826292 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c824x\" (UniqueName: \"kubernetes.io/projected/531c7e95-83b1-4ad1-ad0e-490d9520e8d6-kube-api-access-c824x\") pod \"node-resolver-p545z\" (UID: \"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\") " pod="openshift-dns/node-resolver-p545z" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826312 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9385a5c9-a156-460a-8ca1-9b342ab835c9-mcd-auth-proxy-config\") pod \"machine-config-daemon-jc7vx\" (UID: \"9385a5c9-a156-460a-8ca1-9b342ab835c9\") " pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826356 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826389 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwhc8\" (UniqueName: \"kubernetes.io/projected/9385a5c9-a156-460a-8ca1-9b342ab835c9-kube-api-access-hwhc8\") pod \"machine-config-daemon-jc7vx\" (UID: \"9385a5c9-a156-460a-8ca1-9b342ab835c9\") " pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826440 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826449 4884 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826459 4884 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826470 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826478 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826486 4884 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826494 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826503 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826512 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826521 4884 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826528 4884 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826536 4884 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826545 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826553 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826563 4884 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826571 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826580 4884 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826589 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826608 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826616 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826624 4884 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826633 4884 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826641 4884 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826649 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826658 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826678 4884 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826686 4884 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826694 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826703 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826711 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826720 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826728 4884 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826754 4884 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826763 4884 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826774 4884 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826782 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826791 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826810 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826819 4884 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826828 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826837 4884 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826846 4884 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826854 4884 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826863 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826872 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826879 4884 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826889 4884 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826898 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826907 4884 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826914 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826922 4884 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826930 4884 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826947 4884 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826964 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826972 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826980 4884 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826988 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.826997 4884 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827005 4884 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827013 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827022 4884 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827030 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827038 4884 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827046 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827054 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827062 4884 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827070 4884 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827078 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827086 4884 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827094 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827104 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827114 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827122 4884 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827130 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827139 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827148 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827166 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827176 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827183 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827191 4884 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827199 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827207 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827215 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827223 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827232 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827240 4884 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827248 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827256 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827265 4884 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827273 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827285 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827293 4884 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827301 4884 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827308 4884 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827316 4884 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827325 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827334 4884 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827342 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827350 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827358 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827366 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827374 4884 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827382 4884 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827391 4884 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827400 4884 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827408 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827417 4884 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827425 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827433 4884 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827441 4884 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827449 4884 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827458 4884 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827466 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827473 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827482 4884 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827490 4884 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827498 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827506 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827514 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827522 4884 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827532 4884 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827540 4884 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827549 4884 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827558 4884 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827569 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827587 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827596 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827605 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827613 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827622 4884 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827630 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827638 4884 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827646 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827655 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827663 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827671 4884 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827683 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827691 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827699 4884 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827707 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827716 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827724 4884 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827732 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827753 4884 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827763 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827771 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827779 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827788 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827796 4884 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827804 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827812 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827820 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827828 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827836 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827845 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827853 4884 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827861 4884 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827868 4884 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827876 4884 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827883 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.827974 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.828074 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.828116 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.831929 4884 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:36:42 crc kubenswrapper[4884]: E1202 01:36:42.832228 4884 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.832362 4884 scope.go:117] "RemoveContainer" containerID="aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.839051 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.869829 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.875884 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.882821 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.888398 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.919925 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.930136 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwhc8\" (UniqueName: \"kubernetes.io/projected/9385a5c9-a156-460a-8ca1-9b342ab835c9-kube-api-access-hwhc8\") pod \"machine-config-daemon-jc7vx\" (UID: \"9385a5c9-a156-460a-8ca1-9b342ab835c9\") " pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.930175 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/9385a5c9-a156-460a-8ca1-9b342ab835c9-rootfs\") pod \"machine-config-daemon-jc7vx\" (UID: \"9385a5c9-a156-460a-8ca1-9b342ab835c9\") " pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.930197 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9385a5c9-a156-460a-8ca1-9b342ab835c9-proxy-tls\") pod \"machine-config-daemon-jc7vx\" (UID: \"9385a5c9-a156-460a-8ca1-9b342ab835c9\") " pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.930218 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/531c7e95-83b1-4ad1-ad0e-490d9520e8d6-hosts-file\") pod \"node-resolver-p545z\" (UID: \"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\") " pod="openshift-dns/node-resolver-p545z" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.930236 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c824x\" (UniqueName: \"kubernetes.io/projected/531c7e95-83b1-4ad1-ad0e-490d9520e8d6-kube-api-access-c824x\") pod \"node-resolver-p545z\" (UID: \"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\") " pod="openshift-dns/node-resolver-p545z" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.930255 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9385a5c9-a156-460a-8ca1-9b342ab835c9-mcd-auth-proxy-config\") pod \"machine-config-daemon-jc7vx\" (UID: \"9385a5c9-a156-460a-8ca1-9b342ab835c9\") " pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.930278 4884 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.930921 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9385a5c9-a156-460a-8ca1-9b342ab835c9-mcd-auth-proxy-config\") pod \"machine-config-daemon-jc7vx\" (UID: \"9385a5c9-a156-460a-8ca1-9b342ab835c9\") " pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.930983 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/531c7e95-83b1-4ad1-ad0e-490d9520e8d6-hosts-file\") pod \"node-resolver-p545z\" (UID: \"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\") " pod="openshift-dns/node-resolver-p545z" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.931284 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/9385a5c9-a156-460a-8ca1-9b342ab835c9-rootfs\") pod \"machine-config-daemon-jc7vx\" (UID: \"9385a5c9-a156-460a-8ca1-9b342ab835c9\") " pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.937139 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.938424 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9385a5c9-a156-460a-8ca1-9b342ab835c9-proxy-tls\") pod \"machine-config-daemon-jc7vx\" (UID: \"9385a5c9-a156-460a-8ca1-9b342ab835c9\") " pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.954584 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwhc8\" (UniqueName: \"kubernetes.io/projected/9385a5c9-a156-460a-8ca1-9b342ab835c9-kube-api-access-hwhc8\") pod \"machine-config-daemon-jc7vx\" (UID: \"9385a5c9-a156-460a-8ca1-9b342ab835c9\") " pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.958189 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.958570 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c824x\" (UniqueName: \"kubernetes.io/projected/531c7e95-83b1-4ad1-ad0e-490d9520e8d6-kube-api-access-c824x\") pod \"node-resolver-p545z\" (UID: \"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\") " pod="openshift-dns/node-resolver-p545z" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.962991 4884 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.972897 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.975022 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.978979 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-p545z" Dec 02 01:36:42 crc kubenswrapper[4884]: I1202 01:36:42.988452 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:42 crc kubenswrapper[4884]: W1202 01:36:42.997318 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9385a5c9_a156_460a_8ca1_9b342ab835c9.slice/crio-f8f19e6cf74ae2a51bdad40aa5a9f6977c0886713ace737a3ab69c129a786188 WatchSource:0}: Error finding container f8f19e6cf74ae2a51bdad40aa5a9f6977c0886713ace737a3ab69c129a786188: Status 404 returned error can't find the container with id f8f19e6cf74ae2a51bdad40aa5a9f6977c0886713ace737a3ab69c129a786188 Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.002824 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.020215 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.030253 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.041138 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c74x8"] Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.041869 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: W1202 01:36:43.044215 4884 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": failed to list *v1.Secret: secrets "ovn-kubernetes-node-dockercfg-pwtwl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Dec 02 01:36:43 crc kubenswrapper[4884]: E1202 01:36:43.045108 4884 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-pwtwl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-node-dockercfg-pwtwl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.047124 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-mlkwv"] Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.047212 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.047245 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.047390 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.047437 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.047288 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.047359 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.047405 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.047733 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-6l592"] Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.047969 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.048476 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.050278 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.054062 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.054302 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.054430 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.054461 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.054601 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.058231 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.062148 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.073802 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.084345 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.096600 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.106538 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.118480 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.129136 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.131484 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-systemd\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.131583 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx2hb\" (UniqueName: \"kubernetes.io/projected/d8c50c39-2888-444e-bf55-49b091559978-kube-api-access-bx2hb\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.131627 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/32059fa0-1cad-40d4-92ac-2ee9e89ed661-os-release\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.131700 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-multus-conf-dir\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132220 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-run-multus-certs\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132265 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-etc-openvswitch\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132297 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-run-ovn-kubernetes\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132321 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-cni-netd\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132389 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/791af0d6-a201-430f-b09d-02c24a4b7503-cni-binary-copy\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132421 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d8c50c39-2888-444e-bf55-49b091559978-ovn-node-metrics-cert\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132464 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-var-lib-kubelet\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132496 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-ovn\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132522 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-node-log\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132549 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-ovnkube-config\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132605 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-log-socket\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132639 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-multus-socket-dir-parent\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132673 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-run-k8s-cni-cncf-io\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132703 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-slash\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132733 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-run-netns\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132811 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-run-netns\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132848 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/32059fa0-1cad-40d4-92ac-2ee9e89ed661-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132876 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7xcl\" (UniqueName: \"kubernetes.io/projected/791af0d6-a201-430f-b09d-02c24a4b7503-kube-api-access-c7xcl\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132907 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-multus-cni-dir\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132935 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/32059fa0-1cad-40d4-92ac-2ee9e89ed661-cni-binary-copy\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132964 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-cnibin\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.132992 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-systemd-units\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133019 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-var-lib-openvswitch\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133046 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-hostroot\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133073 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-etc-kubernetes\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133102 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-ovnkube-script-lib\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133165 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghczf\" (UniqueName: \"kubernetes.io/projected/32059fa0-1cad-40d4-92ac-2ee9e89ed661-kube-api-access-ghczf\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133211 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-os-release\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133243 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-openvswitch\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133268 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-cni-bin\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133295 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-var-lib-cni-multus\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133325 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/791af0d6-a201-430f-b09d-02c24a4b7503-multus-daemon-config\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133351 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-env-overrides\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133379 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-system-cni-dir\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133411 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-kubelet\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133439 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/32059fa0-1cad-40d4-92ac-2ee9e89ed661-system-cni-dir\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133464 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-var-lib-cni-bin\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133498 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133525 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/32059fa0-1cad-40d4-92ac-2ee9e89ed661-cnibin\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.133554 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/32059fa0-1cad-40d4-92ac-2ee9e89ed661-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.139378 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.152493 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.162906 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.180835 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.191467 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.206248 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.216429 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.226325 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234398 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234490 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/791af0d6-a201-430f-b09d-02c24a4b7503-cni-binary-copy\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234510 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d8c50c39-2888-444e-bf55-49b091559978-ovn-node-metrics-cert\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234534 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-var-lib-kubelet\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234549 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-ovn\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234563 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-node-log\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234580 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-ovnkube-config\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234609 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-log-socket\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234630 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-multus-socket-dir-parent\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234645 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-run-k8s-cni-cncf-io\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234662 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-slash\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234676 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-run-netns\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234690 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-run-netns\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234713 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/32059fa0-1cad-40d4-92ac-2ee9e89ed661-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234730 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7xcl\" (UniqueName: \"kubernetes.io/projected/791af0d6-a201-430f-b09d-02c24a4b7503-kube-api-access-c7xcl\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234763 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-multus-cni-dir\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234780 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/32059fa0-1cad-40d4-92ac-2ee9e89ed661-cni-binary-copy\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234798 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-cnibin\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234815 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-systemd-units\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234832 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-var-lib-openvswitch\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234846 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-hostroot\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234861 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-etc-kubernetes\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234879 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-ovnkube-script-lib\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234894 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghczf\" (UniqueName: \"kubernetes.io/projected/32059fa0-1cad-40d4-92ac-2ee9e89ed661-kube-api-access-ghczf\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234914 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-os-release\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234944 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-openvswitch\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234960 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-cni-bin\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234977 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-var-lib-cni-multus\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.234992 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/791af0d6-a201-430f-b09d-02c24a4b7503-multus-daemon-config\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235010 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-env-overrides\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235027 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-system-cni-dir\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235041 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-kubelet\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235056 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/32059fa0-1cad-40d4-92ac-2ee9e89ed661-system-cni-dir\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235074 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-var-lib-cni-bin\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235091 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235107 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/32059fa0-1cad-40d4-92ac-2ee9e89ed661-cnibin\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235124 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/32059fa0-1cad-40d4-92ac-2ee9e89ed661-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235140 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-systemd\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235154 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx2hb\" (UniqueName: \"kubernetes.io/projected/d8c50c39-2888-444e-bf55-49b091559978-kube-api-access-bx2hb\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235168 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/32059fa0-1cad-40d4-92ac-2ee9e89ed661-os-release\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235184 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-multus-conf-dir\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235198 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-run-multus-certs\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235211 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-etc-openvswitch\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235227 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-run-ovn-kubernetes\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235248 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-cni-netd\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235310 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-cni-netd\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: E1202 01:36:43.235383 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:36:44.235368418 +0000 UTC m=+20.911205292 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.235971 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/791af0d6-a201-430f-b09d-02c24a4b7503-cni-binary-copy\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.236882 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-run-netns\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.236958 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-var-lib-cni-multus\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.236941 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-cni-bin\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.236979 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-cnibin\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237046 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.236916 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-openvswitch\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237165 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-ovn\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237203 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-run-k8s-cni-cncf-io\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237157 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-run-ovn-kubernetes\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237189 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-run-multus-certs\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237235 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-etc-openvswitch\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237232 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-run-netns\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237282 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-var-lib-cni-bin\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237286 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/32059fa0-1cad-40d4-92ac-2ee9e89ed661-system-cni-dir\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237307 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-log-socket\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237266 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/32059fa0-1cad-40d4-92ac-2ee9e89ed661-cnibin\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237239 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-os-release\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237245 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/32059fa0-1cad-40d4-92ac-2ee9e89ed661-os-release\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237251 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-slash\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237346 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-multus-socket-dir-parent\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237387 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-node-log\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237414 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-host-var-lib-kubelet\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237472 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-multus-cni-dir\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237498 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-hostroot\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237714 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-multus-conf-dir\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237860 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-ovnkube-script-lib\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237869 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/32059fa0-1cad-40d4-92ac-2ee9e89ed661-cni-binary-copy\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237907 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-etc-kubernetes\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237940 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-systemd\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237967 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-kubelet\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.237998 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-systemd-units\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.238026 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-var-lib-openvswitch\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.238119 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/791af0d6-a201-430f-b09d-02c24a4b7503-multus-daemon-config\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.238264 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/791af0d6-a201-430f-b09d-02c24a4b7503-system-cni-dir\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.238285 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/32059fa0-1cad-40d4-92ac-2ee9e89ed661-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.238329 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/32059fa0-1cad-40d4-92ac-2ee9e89ed661-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.238370 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-ovnkube-config\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.238480 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-env-overrides\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.239392 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d8c50c39-2888-444e-bf55-49b091559978-ovn-node-metrics-cert\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.241519 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.256428 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghczf\" (UniqueName: \"kubernetes.io/projected/32059fa0-1cad-40d4-92ac-2ee9e89ed661-kube-api-access-ghczf\") pod \"multus-additional-cni-plugins-mlkwv\" (UID: \"32059fa0-1cad-40d4-92ac-2ee9e89ed661\") " pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.257104 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7xcl\" (UniqueName: \"kubernetes.io/projected/791af0d6-a201-430f-b09d-02c24a4b7503-kube-api-access-c7xcl\") pod \"multus-6l592\" (UID: \"791af0d6-a201-430f-b09d-02c24a4b7503\") " pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.257102 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.259408 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx2hb\" (UniqueName: \"kubernetes.io/projected/d8c50c39-2888-444e-bf55-49b091559978-kube-api-access-bx2hb\") pod \"ovnkube-node-c74x8\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.267140 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.268680 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.281926 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.283362 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.285298 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.300387 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.309969 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.319467 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.329096 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.336111 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.336163 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.336193 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.336227 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:43 crc kubenswrapper[4884]: E1202 01:36:43.336288 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:36:43 crc kubenswrapper[4884]: E1202 01:36:43.336320 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:36:43 crc kubenswrapper[4884]: E1202 01:36:43.336331 4884 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:43 crc kubenswrapper[4884]: E1202 01:36:43.336355 4884 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:36:43 crc kubenswrapper[4884]: E1202 01:36:43.336382 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:36:43 crc kubenswrapper[4884]: E1202 01:36:43.336413 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:36:43 crc kubenswrapper[4884]: E1202 01:36:43.336426 4884 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:36:43 crc kubenswrapper[4884]: E1202 01:36:43.336428 4884 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:43 crc kubenswrapper[4884]: E1202 01:36:43.336425 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:44.336405168 +0000 UTC m=+21.012242102 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:43 crc kubenswrapper[4884]: E1202 01:36:43.336508 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:44.33649049 +0000 UTC m=+21.012327384 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:36:43 crc kubenswrapper[4884]: E1202 01:36:43.336527 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:44.336517751 +0000 UTC m=+21.012354645 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:36:43 crc kubenswrapper[4884]: E1202 01:36:43.336989 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:44.33690434 +0000 UTC m=+21.012741214 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.338569 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.356809 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.385198 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6l592" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.396597 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 02 01:36:43 crc kubenswrapper[4884]: W1202 01:36:43.400108 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod791af0d6_a201_430f_b09d_02c24a4b7503.slice/crio-d146f7a8b0d677834db9ac77b103a576f549d648d5f4f0b23cdf78cdf63dafc4 WatchSource:0}: Error finding container d146f7a8b0d677834db9ac77b103a576f549d648d5f4f0b23cdf78cdf63dafc4: Status 404 returned error can't find the container with id d146f7a8b0d677834db9ac77b103a576f549d648d5f4f0b23cdf78cdf63dafc4 Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.430883 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.445972 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:43 crc kubenswrapper[4884]: W1202 01:36:43.449518 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32059fa0_1cad_40d4_92ac_2ee9e89ed661.slice/crio-104519040dd6ccf754258974fcd973d7af78139c9da30f7e535939417333cbd4 WatchSource:0}: Error finding container 104519040dd6ccf754258974fcd973d7af78139c9da30f7e535939417333cbd4: Status 404 returned error can't find the container with id 104519040dd6ccf754258974fcd973d7af78139c9da30f7e535939417333cbd4 Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.481730 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.531638 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.563318 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.603260 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.616865 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.617647 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.618760 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.619373 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.620353 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.620908 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.621509 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.622527 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.623190 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.624078 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.624600 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.625640 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.626174 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.626672 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.627634 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.628147 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.629103 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.629522 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.630086 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.631065 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.631525 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.632595 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.633099 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.634170 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.634662 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.635283 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.636445 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.636946 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.637492 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.637905 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.638430 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.639337 4884 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.639434 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.641186 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.642263 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.642801 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.644614 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.645280 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.646363 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.647021 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.648115 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.648663 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.649687 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.650340 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.651276 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.651727 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.652608 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.653268 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.654372 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.654988 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.655839 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.656298 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.657232 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.657827 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.658271 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.680412 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.727727 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.766196 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.800174 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.836804 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.880066 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.916628 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.932535 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.946641 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" event={"ID":"32059fa0-1cad-40d4-92ac-2ee9e89ed661","Type":"ContainerStarted","Data":"104519040dd6ccf754258974fcd973d7af78139c9da30f7e535939417333cbd4"} Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.949581 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6l592" event={"ID":"791af0d6-a201-430f-b09d-02c24a4b7503","Type":"ContainerStarted","Data":"4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95"} Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.949660 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6l592" event={"ID":"791af0d6-a201-430f-b09d-02c24a4b7503","Type":"ContainerStarted","Data":"d146f7a8b0d677834db9ac77b103a576f549d648d5f4f0b23cdf78cdf63dafc4"} Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.953243 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec"} Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.953338 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f"} Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.953365 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"f8f19e6cf74ae2a51bdad40aa5a9f6977c0886713ace737a3ab69c129a786188"} Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.960644 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-p545z" event={"ID":"531c7e95-83b1-4ad1-ad0e-490d9520e8d6","Type":"ContainerStarted","Data":"f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3"} Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.960698 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-p545z" event={"ID":"531c7e95-83b1-4ad1-ad0e-490d9520e8d6","Type":"ContainerStarted","Data":"342e3d95077ad68708e5e9b0b619dcf6594caa11532d7f2c22f4ff2c2cf8446c"} Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.965964 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a"} Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.966012 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec"} Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.966028 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"76fd22885798df5771c042676c0ca53438ecce75f4e6688a977c83d8d157b94f"} Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.967966 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.969666 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895"} Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.969719 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f4224facf2df8d59193051820965e2ed02d53e689233209828ace0650a8dba09"} Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.978058 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.981841 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4"} Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.983166 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:36:43 crc kubenswrapper[4884]: I1202 01:36:43.985853 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1d9725b0ce48917b2de6a53295413d035e553a7ff3e6ff52a3a95ff264f880f0"} Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:43.999996 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.022017 4884 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.064050 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.068945 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.078731 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:44 crc kubenswrapper[4884]: W1202 01:36:44.090790 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8c50c39_2888_444e_bf55_49b091559978.slice/crio-d0f1bec63388ff2bd42e26b0db45f584c2c0ace4e18a79e18be00e8b0493a7cc WatchSource:0}: Error finding container d0f1bec63388ff2bd42e26b0db45f584c2c0ace4e18a79e18be00e8b0493a7cc: Status 404 returned error can't find the container with id d0f1bec63388ff2bd42e26b0db45f584c2c0ace4e18a79e18be00e8b0493a7cc Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.133331 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.166733 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.207237 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.237847 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.253119 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.253488 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:36:46.253465042 +0000 UTC m=+22.929301926 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.282078 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.318209 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.354991 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.355031 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.355051 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.355077 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.355178 4884 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.355232 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.355266 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.355280 4884 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.355241 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:46.35522758 +0000 UTC m=+23.031064464 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.355349 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:46.355333363 +0000 UTC m=+23.031170247 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.355379 4884 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.355403 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:46.355396884 +0000 UTC m=+23.031233768 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.355446 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.355455 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.355462 4884 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.355480 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:46.355474996 +0000 UTC m=+23.031311880 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.369622 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.400765 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.441482 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.485185 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.523203 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.567340 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.605447 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.613998 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.614077 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.614132 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.614231 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.614077 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:44 crc kubenswrapper[4884]: E1202 01:36:44.614344 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.640233 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.680334 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.720409 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.758623 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.806708 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.844698 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.882260 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.924116 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.964318 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:44Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.992368 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8c50c39-2888-444e-bf55-49b091559978" containerID="36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997" exitCode=0 Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.992450 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerDied","Data":"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997"} Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.992543 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerStarted","Data":"d0f1bec63388ff2bd42e26b0db45f584c2c0ace4e18a79e18be00e8b0493a7cc"} Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.995576 4884 generic.go:334] "Generic (PLEG): container finished" podID="32059fa0-1cad-40d4-92ac-2ee9e89ed661" containerID="7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e" exitCode=0 Dec 02 01:36:44 crc kubenswrapper[4884]: I1202 01:36:44.995685 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" event={"ID":"32059fa0-1cad-40d4-92ac-2ee9e89ed661","Type":"ContainerDied","Data":"7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e"} Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.024500 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.047196 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.083699 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.123696 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.159027 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.199468 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.250688 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.281651 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.381994 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.402833 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.426395 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.443818 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.476407 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.517080 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.558683 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.596445 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.637954 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.680224 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.726510 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.759135 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.801258 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.845972 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.886176 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.920385 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:45 crc kubenswrapper[4884]: I1202 01:36:45.960636 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:45Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.001591 4884 generic.go:334] "Generic (PLEG): container finished" podID="32059fa0-1cad-40d4-92ac-2ee9e89ed661" containerID="bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6" exitCode=0 Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.001685 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" event={"ID":"32059fa0-1cad-40d4-92ac-2ee9e89ed661","Type":"ContainerDied","Data":"bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6"} Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.007681 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerStarted","Data":"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055"} Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.007769 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerStarted","Data":"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96"} Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.007791 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerStarted","Data":"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b"} Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.007812 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerStarted","Data":"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518"} Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.007833 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerStarted","Data":"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623"} Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.007852 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerStarted","Data":"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844"} Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.030212 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.050897 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.097316 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.119878 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.160059 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.205235 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.242796 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.277000 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.277221 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:36:50.277204406 +0000 UTC m=+26.953041300 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.284245 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.322731 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.366671 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.377993 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.378067 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.378103 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.378135 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.378185 4884 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.378231 4884 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.378275 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:50.378255697 +0000 UTC m=+27.054092581 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.378297 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:50.378288198 +0000 UTC m=+27.054125082 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.378367 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.378389 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.378407 4884 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.378455 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:50.378436182 +0000 UTC m=+27.054273096 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.378528 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.378546 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.378559 4884 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.378596 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:50.378583565 +0000 UTC m=+27.054420479 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.399637 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.423575 4884 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.425920 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.425965 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.425979 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.426076 4884 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.441960 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.491469 4884 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.492062 4884 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.493618 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.493664 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.493675 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.493694 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.493707 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:46Z","lastTransitionTime":"2025-12-02T01:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.507636 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.511442 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.511498 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.511516 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.511541 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.511559 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:46Z","lastTransitionTime":"2025-12-02T01:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.518933 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.526529 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.530557 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.530594 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.530605 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.530623 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.530637 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:46Z","lastTransitionTime":"2025-12-02T01:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.547205 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.551165 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.551208 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.551221 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.551239 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.551276 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:46Z","lastTransitionTime":"2025-12-02T01:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.563985 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.568137 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.572260 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.572301 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.572313 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.572332 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.572345 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:46Z","lastTransitionTime":"2025-12-02T01:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.585463 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.585614 4884 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.587351 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.587394 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.587407 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.587425 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.587461 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:46Z","lastTransitionTime":"2025-12-02T01:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.612776 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.613230 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.613264 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.613343 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.613396 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.613411 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:36:46 crc kubenswrapper[4884]: E1202 01:36:46.613552 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.650032 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.680556 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:46Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.690511 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.690555 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.690571 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.690599 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.690618 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:46Z","lastTransitionTime":"2025-12-02T01:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.793514 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.793563 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.793582 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.793608 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.793626 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:46Z","lastTransitionTime":"2025-12-02T01:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.896045 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.896078 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.896089 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.896105 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:46 crc kubenswrapper[4884]: I1202 01:36:46.896119 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:46Z","lastTransitionTime":"2025-12-02T01:36:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.002610 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.003380 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.003416 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.003499 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.003521 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:47Z","lastTransitionTime":"2025-12-02T01:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.014264 4884 generic.go:334] "Generic (PLEG): container finished" podID="32059fa0-1cad-40d4-92ac-2ee9e89ed661" containerID="f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f" exitCode=0 Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.014388 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" event={"ID":"32059fa0-1cad-40d4-92ac-2ee9e89ed661","Type":"ContainerDied","Data":"f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f"} Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.016932 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02"} Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.051974 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.066184 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.087338 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.108341 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.108690 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.108724 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.108759 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.108779 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.108792 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:47Z","lastTransitionTime":"2025-12-02T01:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.125561 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.138772 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.153401 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.172659 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.191476 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.208363 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.211233 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.211272 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.211284 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.211299 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.211311 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:47Z","lastTransitionTime":"2025-12-02T01:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.222508 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.237946 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.255349 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.271264 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.296973 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.313714 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.313765 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.313775 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.313789 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.313799 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:47Z","lastTransitionTime":"2025-12-02T01:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.322762 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.366421 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.401982 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.415588 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.415649 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.415667 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.415689 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.415706 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:47Z","lastTransitionTime":"2025-12-02T01:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.445777 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.484138 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.518374 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.518431 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.518450 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.518475 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.518495 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:47Z","lastTransitionTime":"2025-12-02T01:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.525646 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.563305 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.605343 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.623335 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.623416 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.623441 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.623471 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.623495 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:47Z","lastTransitionTime":"2025-12-02T01:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.646432 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.684256 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.726676 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.726783 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.726809 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.726840 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.726862 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:47Z","lastTransitionTime":"2025-12-02T01:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.729234 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.762714 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.802678 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:47Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.830490 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.830591 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.830611 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.830644 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.830667 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:47Z","lastTransitionTime":"2025-12-02T01:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.932624 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.932658 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.932669 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.932686 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:47 crc kubenswrapper[4884]: I1202 01:36:47.932698 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:47Z","lastTransitionTime":"2025-12-02T01:36:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.027645 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerStarted","Data":"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782"} Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.030959 4884 generic.go:334] "Generic (PLEG): container finished" podID="32059fa0-1cad-40d4-92ac-2ee9e89ed661" containerID="0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69" exitCode=0 Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.031053 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" event={"ID":"32059fa0-1cad-40d4-92ac-2ee9e89ed661","Type":"ContainerDied","Data":"0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69"} Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.037120 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.037177 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.037194 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.037218 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.037236 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:48Z","lastTransitionTime":"2025-12-02T01:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.055849 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.075188 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.105564 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.121970 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.136582 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.139634 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.139692 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.139710 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.139736 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.139779 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:48Z","lastTransitionTime":"2025-12-02T01:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.146781 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.160711 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.177715 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.189883 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.203720 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.242943 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.243948 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.244001 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.244018 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.244093 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.244111 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:48Z","lastTransitionTime":"2025-12-02T01:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.280884 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.319136 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.346556 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.346598 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.346611 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.346630 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.346643 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:48Z","lastTransitionTime":"2025-12-02T01:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.362943 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.449657 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.449702 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.449715 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.449736 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.449766 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:48Z","lastTransitionTime":"2025-12-02T01:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.552514 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.552834 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.552991 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.553141 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.553293 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:48Z","lastTransitionTime":"2025-12-02T01:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.613431 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:36:48 crc kubenswrapper[4884]: E1202 01:36:48.613615 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.613662 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.613729 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:36:48 crc kubenswrapper[4884]: E1202 01:36:48.613859 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:36:48 crc kubenswrapper[4884]: E1202 01:36:48.613981 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.657122 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.657188 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.657206 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.657243 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.657265 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:48Z","lastTransitionTime":"2025-12-02T01:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.760219 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.760543 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.760681 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.760859 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.761014 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:48Z","lastTransitionTime":"2025-12-02T01:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.864613 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.864696 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.864713 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.864770 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.865335 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:48Z","lastTransitionTime":"2025-12-02T01:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.969075 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.969129 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.969146 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.969172 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:48 crc kubenswrapper[4884]: I1202 01:36:48.969190 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:48Z","lastTransitionTime":"2025-12-02T01:36:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.039951 4884 generic.go:334] "Generic (PLEG): container finished" podID="32059fa0-1cad-40d4-92ac-2ee9e89ed661" containerID="2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046" exitCode=0 Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.040047 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" event={"ID":"32059fa0-1cad-40d4-92ac-2ee9e89ed661","Type":"ContainerDied","Data":"2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046"} Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.063235 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:49Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.072141 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.072175 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.072185 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.072201 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.072212 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:49Z","lastTransitionTime":"2025-12-02T01:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.079408 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:49Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.096132 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:49Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.127186 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:49Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.151055 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:49Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.178975 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.179017 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.179029 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.179047 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.179059 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:49Z","lastTransitionTime":"2025-12-02T01:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.180392 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:49Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.198843 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:49Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.212112 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:49Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.224219 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:49Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.239692 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:49Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.259288 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:49Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.276006 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:49Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.281540 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.281590 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.281605 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.281621 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.281633 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:49Z","lastTransitionTime":"2025-12-02T01:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.292356 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:49Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.304797 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:49Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.384862 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.384938 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.384961 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.384991 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.385015 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:49Z","lastTransitionTime":"2025-12-02T01:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.488997 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.489059 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.489076 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.489101 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.489118 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:49Z","lastTransitionTime":"2025-12-02T01:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.593326 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.593446 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.593471 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.593503 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.593527 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:49Z","lastTransitionTime":"2025-12-02T01:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.696475 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.696533 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.696549 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.696582 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.696599 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:49Z","lastTransitionTime":"2025-12-02T01:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.798966 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.799011 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.799032 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.799055 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.799072 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:49Z","lastTransitionTime":"2025-12-02T01:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.901816 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.902161 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.902388 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.902554 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:49 crc kubenswrapper[4884]: I1202 01:36:49.902712 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:49Z","lastTransitionTime":"2025-12-02T01:36:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.005900 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.005960 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.005998 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.006022 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.006039 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:50Z","lastTransitionTime":"2025-12-02T01:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.048023 4884 generic.go:334] "Generic (PLEG): container finished" podID="32059fa0-1cad-40d4-92ac-2ee9e89ed661" containerID="84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639" exitCode=0 Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.049896 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" event={"ID":"32059fa0-1cad-40d4-92ac-2ee9e89ed661","Type":"ContainerDied","Data":"84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639"} Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.075956 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:50Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.100304 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:50Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.108650 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.108712 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.108791 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.108833 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.108857 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:50Z","lastTransitionTime":"2025-12-02T01:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.118235 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:50Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.143481 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:50Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.161817 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:50Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.189091 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:50Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.206950 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:50Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.212850 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.212909 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.212928 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.212953 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.212975 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:50Z","lastTransitionTime":"2025-12-02T01:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.226904 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:50Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.243484 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:50Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.262930 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:50Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.282063 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:50Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.300471 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:50Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.314081 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:50Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.317210 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.317250 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.317268 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.317291 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.317309 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:50Z","lastTransitionTime":"2025-12-02T01:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.319614 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.319902 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:36:58.319876559 +0000 UTC m=+34.995713483 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.328400 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:50Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.419976 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.420008 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.420016 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.420031 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.420040 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:50Z","lastTransitionTime":"2025-12-02T01:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.420215 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.420283 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.420320 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.420337 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.420358 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.420371 4884 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.420374 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.420415 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.420438 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.420450 4884 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.420419 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:58.420402638 +0000 UTC m=+35.096239542 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.420466 4884 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.420532 4884 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.420495 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:58.4204849 +0000 UTC m=+35.096321784 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.420631 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:58.420601982 +0000 UTC m=+35.096438906 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.420678 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:36:58.420662644 +0000 UTC m=+35.096499568 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.522692 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.522811 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.522839 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.522868 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.522890 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:50Z","lastTransitionTime":"2025-12-02T01:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.613834 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.613941 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.614066 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.614081 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.614268 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:36:50 crc kubenswrapper[4884]: E1202 01:36:50.614454 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.625555 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.625611 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.625629 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.625654 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.625673 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:50Z","lastTransitionTime":"2025-12-02T01:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.729083 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.729133 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.729154 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.729183 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.729206 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:50Z","lastTransitionTime":"2025-12-02T01:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.832021 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.832081 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.832098 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.832124 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.832141 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:50Z","lastTransitionTime":"2025-12-02T01:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.934784 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.934841 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.934858 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.934886 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:50 crc kubenswrapper[4884]: I1202 01:36:50.934904 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:50Z","lastTransitionTime":"2025-12-02T01:36:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.038567 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.038622 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.038639 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.038663 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.038681 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:51Z","lastTransitionTime":"2025-12-02T01:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.061089 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerStarted","Data":"3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc"} Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.061884 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.061927 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.072031 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" event={"ID":"32059fa0-1cad-40d4-92ac-2ee9e89ed661","Type":"ContainerStarted","Data":"412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6"} Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.086620 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.098130 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.098472 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.110618 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.141233 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.141281 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.141299 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.141322 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.141338 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:51Z","lastTransitionTime":"2025-12-02T01:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.142629 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.160284 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.177162 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.190590 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.210865 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.234353 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.244588 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.244772 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.244867 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.244945 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.245022 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:51Z","lastTransitionTime":"2025-12-02T01:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.258284 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.279422 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.294553 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.311140 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.330982 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.347095 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.347161 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.347179 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.347204 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.347222 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:51Z","lastTransitionTime":"2025-12-02T01:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.357191 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.377037 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.392580 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.408237 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.444654 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.450325 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.450394 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.450423 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.450455 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.450479 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:51Z","lastTransitionTime":"2025-12-02T01:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.461815 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.487797 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.509325 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.529396 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.550773 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.553029 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.553085 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.553103 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.553130 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.553147 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:51Z","lastTransitionTime":"2025-12-02T01:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.568665 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.586425 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.608147 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.628935 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.649473 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:51Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.655656 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.655870 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.655996 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.656114 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.656276 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:51Z","lastTransitionTime":"2025-12-02T01:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.758725 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.758829 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.758850 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.758883 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.758900 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:51Z","lastTransitionTime":"2025-12-02T01:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.865157 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.865286 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.865306 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.865332 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.865347 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:51Z","lastTransitionTime":"2025-12-02T01:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.968698 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.968890 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.968916 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.968942 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:51 crc kubenswrapper[4884]: I1202 01:36:51.968960 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:51Z","lastTransitionTime":"2025-12-02T01:36:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.072606 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.072669 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.072696 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.072726 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.072782 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:52Z","lastTransitionTime":"2025-12-02T01:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.075007 4884 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.175724 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.175997 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.176097 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.176188 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.176264 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:52Z","lastTransitionTime":"2025-12-02T01:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.279810 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.280043 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.280053 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.280069 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.280080 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:52Z","lastTransitionTime":"2025-12-02T01:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.383164 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.383216 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.383230 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.383249 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.383264 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:52Z","lastTransitionTime":"2025-12-02T01:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.485800 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.486069 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.486150 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.486263 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.486388 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:52Z","lastTransitionTime":"2025-12-02T01:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.589058 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.589114 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.589128 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.589144 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.589157 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:52Z","lastTransitionTime":"2025-12-02T01:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.613807 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:36:52 crc kubenswrapper[4884]: E1202 01:36:52.613967 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.614095 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:36:52 crc kubenswrapper[4884]: E1202 01:36:52.614290 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.613826 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:52 crc kubenswrapper[4884]: E1202 01:36:52.614521 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.691713 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.692022 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.692134 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.692237 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.692332 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:52Z","lastTransitionTime":"2025-12-02T01:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.795453 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.795505 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.795517 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.795534 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.795545 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:52Z","lastTransitionTime":"2025-12-02T01:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.898762 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.898817 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.898828 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.898844 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:52 crc kubenswrapper[4884]: I1202 01:36:52.898855 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:52Z","lastTransitionTime":"2025-12-02T01:36:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.002388 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.002443 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.002459 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.002485 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.002513 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:53Z","lastTransitionTime":"2025-12-02T01:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.079546 4884 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.105999 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.106098 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.106116 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.106144 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.106162 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:53Z","lastTransitionTime":"2025-12-02T01:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.209510 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.209560 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.209578 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.209602 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.209620 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:53Z","lastTransitionTime":"2025-12-02T01:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.313087 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.313283 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.313650 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.314012 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.314047 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:53Z","lastTransitionTime":"2025-12-02T01:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.417228 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.417292 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.417312 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.417338 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.417356 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:53Z","lastTransitionTime":"2025-12-02T01:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.521060 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.521152 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.521174 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.521192 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.521206 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:53Z","lastTransitionTime":"2025-12-02T01:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.628620 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.628683 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.628704 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.628734 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.628782 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:53Z","lastTransitionTime":"2025-12-02T01:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.640621 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.664264 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.679761 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.693417 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.712731 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.727472 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.732084 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.732145 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.732170 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.732200 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.732223 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:53Z","lastTransitionTime":"2025-12-02T01:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.750979 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.771817 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.803361 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.819998 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.834683 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.834787 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.834807 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.834834 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.834857 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:53Z","lastTransitionTime":"2025-12-02T01:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.836424 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.855702 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.873234 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.891938 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.939565 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.939631 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.939655 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.939687 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:53 crc kubenswrapper[4884]: I1202 01:36:53.939710 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:53Z","lastTransitionTime":"2025-12-02T01:36:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.042175 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.042219 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.042229 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.042244 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.042254 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:54Z","lastTransitionTime":"2025-12-02T01:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.088746 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/0.log" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.094305 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8c50c39-2888-444e-bf55-49b091559978" containerID="3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc" exitCode=1 Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.094352 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerDied","Data":"3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc"} Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.095139 4884 scope.go:117] "RemoveContainer" containerID="3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.117399 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.139147 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.147844 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.147875 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.147902 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.147918 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.147927 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:54Z","lastTransitionTime":"2025-12-02T01:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.160289 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.178959 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.199050 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.212613 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.248981 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.251825 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.251900 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.251927 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.251961 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.251987 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:54Z","lastTransitionTime":"2025-12-02T01:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.264783 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.295159 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:53Z\\\",\\\"message\\\":\\\":140\\\\nI1202 01:36:53.308600 6157 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1202 01:36:53.304048 6157 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:53.308985 6157 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1202 01:36:53.309005 6157 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1202 01:36:53.308726 6157 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:53.308810 6157 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1202 01:36:53.309424 6157 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:53.309951 6157 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1202 01:36:53.310038 6157 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:53.310130 6157 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:53.310144 6157 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1202 01:36:53.310205 6157 factory.go:656] Stopping watch factory\\\\nI1202 01:36:53.310238 6157 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:53.310233 6157 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1202 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.313345 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.330361 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.345964 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.354870 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.354911 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.354927 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.354949 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.354966 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:54Z","lastTransitionTime":"2025-12-02T01:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.367139 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.384795 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.457869 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.457925 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.457943 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.457969 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.457987 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:54Z","lastTransitionTime":"2025-12-02T01:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.560418 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.560518 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.560541 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.560808 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.560847 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:54Z","lastTransitionTime":"2025-12-02T01:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.614069 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.614108 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.614083 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:36:54 crc kubenswrapper[4884]: E1202 01:36:54.614265 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:36:54 crc kubenswrapper[4884]: E1202 01:36:54.614500 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:36:54 crc kubenswrapper[4884]: E1202 01:36:54.614646 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.663411 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.663447 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.663455 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.663468 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.663477 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:54Z","lastTransitionTime":"2025-12-02T01:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.766549 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.766619 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.766642 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.766672 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.766694 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:54Z","lastTransitionTime":"2025-12-02T01:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.868967 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.869014 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.869032 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.869057 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.869077 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:54Z","lastTransitionTime":"2025-12-02T01:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.926286 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4"] Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.927117 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.932901 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.933397 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.956198 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.968806 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ec4bbba9-9a8c-489e-8a21-198f1cb12a57-env-overrides\") pod \"ovnkube-control-plane-749d76644c-fdrt4\" (UID: \"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.968860 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pln92\" (UniqueName: \"kubernetes.io/projected/ec4bbba9-9a8c-489e-8a21-198f1cb12a57-kube-api-access-pln92\") pod \"ovnkube-control-plane-749d76644c-fdrt4\" (UID: \"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.968945 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ec4bbba9-9a8c-489e-8a21-198f1cb12a57-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-fdrt4\" (UID: \"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.968997 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ec4bbba9-9a8c-489e-8a21-198f1cb12a57-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-fdrt4\" (UID: \"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.972231 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.972287 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.972305 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.972329 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.972349 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:54Z","lastTransitionTime":"2025-12-02T01:36:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.977404 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:54 crc kubenswrapper[4884]: I1202 01:36:54.994163 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:54Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.015334 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.038917 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.061063 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.070050 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ec4bbba9-9a8c-489e-8a21-198f1cb12a57-env-overrides\") pod \"ovnkube-control-plane-749d76644c-fdrt4\" (UID: \"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.070305 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pln92\" (UniqueName: \"kubernetes.io/projected/ec4bbba9-9a8c-489e-8a21-198f1cb12a57-kube-api-access-pln92\") pod \"ovnkube-control-plane-749d76644c-fdrt4\" (UID: \"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.070565 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ec4bbba9-9a8c-489e-8a21-198f1cb12a57-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-fdrt4\" (UID: \"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.070733 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ec4bbba9-9a8c-489e-8a21-198f1cb12a57-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-fdrt4\" (UID: \"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.071122 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ec4bbba9-9a8c-489e-8a21-198f1cb12a57-env-overrides\") pod \"ovnkube-control-plane-749d76644c-fdrt4\" (UID: \"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.071826 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ec4bbba9-9a8c-489e-8a21-198f1cb12a57-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-fdrt4\" (UID: \"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.075004 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.075068 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.075085 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.075109 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.075128 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:55Z","lastTransitionTime":"2025-12-02T01:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.083116 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.083914 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ec4bbba9-9a8c-489e-8a21-198f1cb12a57-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-fdrt4\" (UID: \"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.101453 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/0.log" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.104413 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pln92\" (UniqueName: \"kubernetes.io/projected/ec4bbba9-9a8c-489e-8a21-198f1cb12a57-kube-api-access-pln92\") pod \"ovnkube-control-plane-749d76644c-fdrt4\" (UID: \"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.111570 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerStarted","Data":"f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c"} Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.112955 4884 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.114371 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.136824 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.154832 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.179075 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.179310 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.179499 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.179602 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.179681 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:55Z","lastTransitionTime":"2025-12-02T01:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.180866 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.197243 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.225818 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.250161 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.252283 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.276945 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:53Z\\\",\\\"message\\\":\\\":140\\\\nI1202 01:36:53.308600 6157 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1202 01:36:53.304048 6157 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:53.308985 6157 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1202 01:36:53.309005 6157 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1202 01:36:53.308726 6157 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:53.308810 6157 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1202 01:36:53.309424 6157 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:53.309951 6157 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1202 01:36:53.310038 6157 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:53.310130 6157 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:53.310144 6157 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1202 01:36:53.310205 6157 factory.go:656] Stopping watch factory\\\\nI1202 01:36:53.310238 6157 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:53.310233 6157 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1202 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.281856 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.281881 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.281889 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.281901 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.281909 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:55Z","lastTransitionTime":"2025-12-02T01:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.303289 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.317112 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.335973 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:53Z\\\",\\\"message\\\":\\\":140\\\\nI1202 01:36:53.308600 6157 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1202 01:36:53.304048 6157 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:53.308985 6157 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1202 01:36:53.309005 6157 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1202 01:36:53.308726 6157 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:53.308810 6157 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1202 01:36:53.309424 6157 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:53.309951 6157 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1202 01:36:53.310038 6157 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:53.310130 6157 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:53.310144 6157 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1202 01:36:53.310205 6157 factory.go:656] Stopping watch factory\\\\nI1202 01:36:53.310238 6157 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:53.310233 6157 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1202 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.352408 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.364797 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.385036 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.385085 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.385102 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.385130 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.385148 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:55Z","lastTransitionTime":"2025-12-02T01:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.402576 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.420547 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.440724 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.453602 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.465627 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.476972 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.487116 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.487140 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.487147 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.487160 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.487169 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:55Z","lastTransitionTime":"2025-12-02T01:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.487627 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.498628 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.518712 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.528280 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:55Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.589700 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.589742 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.589769 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.589787 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.589800 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:55Z","lastTransitionTime":"2025-12-02T01:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.692461 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.692524 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.692543 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.692568 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.692586 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:55Z","lastTransitionTime":"2025-12-02T01:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.794863 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.794898 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.794911 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.794929 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.794941 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:55Z","lastTransitionTime":"2025-12-02T01:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.897465 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.897511 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.897526 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.897545 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:55 crc kubenswrapper[4884]: I1202 01:36:55.897560 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:55Z","lastTransitionTime":"2025-12-02T01:36:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.000647 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.000727 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.000791 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.000825 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.000849 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:56Z","lastTransitionTime":"2025-12-02T01:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.104381 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.104443 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.104460 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.104491 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.104512 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:56Z","lastTransitionTime":"2025-12-02T01:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.119632 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" event={"ID":"ec4bbba9-9a8c-489e-8a21-198f1cb12a57","Type":"ContainerStarted","Data":"1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.119707 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" event={"ID":"ec4bbba9-9a8c-489e-8a21-198f1cb12a57","Type":"ContainerStarted","Data":"59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.119728 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" event={"ID":"ec4bbba9-9a8c-489e-8a21-198f1cb12a57","Type":"ContainerStarted","Data":"33a8623ce82f823b749adeb760282e345702b0abfee66153670daf883b594481"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.126180 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/1.log" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.127510 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/0.log" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.135229 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8c50c39-2888-444e-bf55-49b091559978" containerID="f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c" exitCode=1 Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.135351 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerDied","Data":"f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.135435 4884 scope.go:117] "RemoveContainer" containerID="3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.140940 4884 scope.go:117] "RemoveContainer" containerID="f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c" Dec 02 01:36:56 crc kubenswrapper[4884]: E1202 01:36:56.141407 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.145584 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.171454 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.195987 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.207238 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.207299 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.207321 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.207351 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.207374 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:56Z","lastTransitionTime":"2025-12-02T01:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.226374 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.243393 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.267853 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:53Z\\\",\\\"message\\\":\\\":140\\\\nI1202 01:36:53.308600 6157 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1202 01:36:53.304048 6157 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:53.308985 6157 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1202 01:36:53.309005 6157 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1202 01:36:53.308726 6157 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:53.308810 6157 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1202 01:36:53.309424 6157 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:53.309951 6157 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1202 01:36:53.310038 6157 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:53.310130 6157 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:53.310144 6157 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1202 01:36:53.310205 6157 factory.go:656] Stopping watch factory\\\\nI1202 01:36:53.310238 6157 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:53.310233 6157 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1202 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.279304 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.291304 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.304406 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.310006 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.310046 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.310055 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.310070 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.310080 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:56Z","lastTransitionTime":"2025-12-02T01:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.318135 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.329619 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.340933 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.359883 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.377472 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.392409 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.406272 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.412282 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.412312 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.412323 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.412351 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.412362 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:56Z","lastTransitionTime":"2025-12-02T01:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.418810 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.439307 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.453711 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.466582 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.478419 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.496895 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.509463 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.514664 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.514712 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.514728 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.514786 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.514805 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:56Z","lastTransitionTime":"2025-12-02T01:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.529188 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.543137 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.560393 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.576440 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.607567 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.613635 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.613669 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:56 crc kubenswrapper[4884]: E1202 01:36:56.613812 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.613827 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:36:56 crc kubenswrapper[4884]: E1202 01:36:56.613923 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:36:56 crc kubenswrapper[4884]: E1202 01:36:56.614011 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.617568 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.617595 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.617607 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.617625 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.617638 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:56Z","lastTransitionTime":"2025-12-02T01:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.628836 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.650723 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:53Z\\\",\\\"message\\\":\\\":140\\\\nI1202 01:36:53.308600 6157 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1202 01:36:53.304048 6157 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:53.308985 6157 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1202 01:36:53.309005 6157 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1202 01:36:53.308726 6157 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:53.308810 6157 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1202 01:36:53.309424 6157 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:53.309951 6157 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1202 01:36:53.310038 6157 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:53.310130 6157 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:53.310144 6157 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1202 01:36:53.310205 6157 factory.go:656] Stopping watch factory\\\\nI1202 01:36:53.310238 6157 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:53.310233 6157 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1202 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:55Z\\\",\\\"message\\\":\\\"t handler 5\\\\nI1202 01:36:55.706399 6304 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:55.706406 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1202 01:36:55.706225 6304 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706522 6304 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:55.706545 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1202 01:36:55.706553 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:55.706575 6304 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706849 6304 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:55.707271 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:55.707390 6304 factory.go:656] Stopping watch factory\\\\nI1202 01:36:55.707426 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.719890 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.719950 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.719968 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.719992 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.720010 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:56Z","lastTransitionTime":"2025-12-02T01:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.822647 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.822735 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.822789 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.822814 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.822832 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:56Z","lastTransitionTime":"2025-12-02T01:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.861306 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-whh6t"] Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.862095 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-ww4r4"] Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.862344 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:36:56 crc kubenswrapper[4884]: E1202 01:36:56.862520 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.862610 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-ww4r4" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.867381 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.867538 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.867683 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.867929 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.888873 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs\") pod \"network-metrics-daemon-whh6t\" (UID: \"7a6a67e3-3b98-45df-9573-6efee410a90c\") " pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.888932 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wtwc\" (UniqueName: \"kubernetes.io/projected/7a6a67e3-3b98-45df-9573-6efee410a90c-kube-api-access-8wtwc\") pod \"network-metrics-daemon-whh6t\" (UID: \"7a6a67e3-3b98-45df-9573-6efee410a90c\") " pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.888976 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f6b8c081-9419-49d2-a77f-c4fc1f136379-serviceca\") pod \"node-ca-ww4r4\" (UID: \"f6b8c081-9419-49d2-a77f-c4fc1f136379\") " pod="openshift-image-registry/node-ca-ww4r4" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.889052 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnr86\" (UniqueName: \"kubernetes.io/projected/f6b8c081-9419-49d2-a77f-c4fc1f136379-kube-api-access-rnr86\") pod \"node-ca-ww4r4\" (UID: \"f6b8c081-9419-49d2-a77f-c4fc1f136379\") " pod="openshift-image-registry/node-ca-ww4r4" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.889092 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6b8c081-9419-49d2-a77f-c4fc1f136379-host\") pod \"node-ca-ww4r4\" (UID: \"f6b8c081-9419-49d2-a77f-c4fc1f136379\") " pod="openshift-image-registry/node-ca-ww4r4" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.894132 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.914511 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.925981 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.926031 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.926049 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.926075 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.926094 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:56Z","lastTransitionTime":"2025-12-02T01:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.932975 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.952416 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.976416 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.976477 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.976495 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.976521 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.976538 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:56Z","lastTransitionTime":"2025-12-02T01:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.980304 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.990510 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs\") pod \"network-metrics-daemon-whh6t\" (UID: \"7a6a67e3-3b98-45df-9573-6efee410a90c\") " pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:36:56 crc kubenswrapper[4884]: E1202 01:36:56.990667 4884 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:36:56 crc kubenswrapper[4884]: E1202 01:36:56.990840 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs podName:7a6a67e3-3b98-45df-9573-6efee410a90c nodeName:}" failed. No retries permitted until 2025-12-02 01:36:57.490818019 +0000 UTC m=+34.166654933 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs") pod "network-metrics-daemon-whh6t" (UID: "7a6a67e3-3b98-45df-9573-6efee410a90c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.991106 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wtwc\" (UniqueName: \"kubernetes.io/projected/7a6a67e3-3b98-45df-9573-6efee410a90c-kube-api-access-8wtwc\") pod \"network-metrics-daemon-whh6t\" (UID: \"7a6a67e3-3b98-45df-9573-6efee410a90c\") " pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.991225 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f6b8c081-9419-49d2-a77f-c4fc1f136379-serviceca\") pod \"node-ca-ww4r4\" (UID: \"f6b8c081-9419-49d2-a77f-c4fc1f136379\") " pod="openshift-image-registry/node-ca-ww4r4" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.991319 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnr86\" (UniqueName: \"kubernetes.io/projected/f6b8c081-9419-49d2-a77f-c4fc1f136379-kube-api-access-rnr86\") pod \"node-ca-ww4r4\" (UID: \"f6b8c081-9419-49d2-a77f-c4fc1f136379\") " pod="openshift-image-registry/node-ca-ww4r4" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.991670 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6b8c081-9419-49d2-a77f-c4fc1f136379-host\") pod \"node-ca-ww4r4\" (UID: \"f6b8c081-9419-49d2-a77f-c4fc1f136379\") " pod="openshift-image-registry/node-ca-ww4r4" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.991784 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f6b8c081-9419-49d2-a77f-c4fc1f136379-host\") pod \"node-ca-ww4r4\" (UID: \"f6b8c081-9419-49d2-a77f-c4fc1f136379\") " pod="openshift-image-registry/node-ca-ww4r4" Dec 02 01:36:56 crc kubenswrapper[4884]: I1202 01:36:56.993131 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/f6b8c081-9419-49d2-a77f-c4fc1f136379-serviceca\") pod \"node-ca-ww4r4\" (UID: \"f6b8c081-9419-49d2-a77f-c4fc1f136379\") " pod="openshift-image-registry/node-ca-ww4r4" Dec 02 01:36:56 crc kubenswrapper[4884]: E1202 01:36:56.996278 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:56Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.000333 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.000389 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.000410 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.000435 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.000488 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:57Z","lastTransitionTime":"2025-12-02T01:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.003175 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.014903 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wtwc\" (UniqueName: \"kubernetes.io/projected/7a6a67e3-3b98-45df-9573-6efee410a90c-kube-api-access-8wtwc\") pod \"network-metrics-daemon-whh6t\" (UID: \"7a6a67e3-3b98-45df-9573-6efee410a90c\") " pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:36:57 crc kubenswrapper[4884]: E1202 01:36:57.021420 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.024691 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.027128 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.027170 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.027188 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.027211 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.027229 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:57Z","lastTransitionTime":"2025-12-02T01:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.028360 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnr86\" (UniqueName: \"kubernetes.io/projected/f6b8c081-9419-49d2-a77f-c4fc1f136379-kube-api-access-rnr86\") pod \"node-ca-ww4r4\" (UID: \"f6b8c081-9419-49d2-a77f-c4fc1f136379\") " pod="openshift-image-registry/node-ca-ww4r4" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.041990 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: E1202 01:36:57.044779 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.049883 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.049943 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.049963 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.049989 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.050008 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:57Z","lastTransitionTime":"2025-12-02T01:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.057193 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: E1202 01:36:57.064171 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.068312 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.068356 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.068368 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.068387 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.068399 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:57Z","lastTransitionTime":"2025-12-02T01:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.071973 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: E1202 01:36:57.081903 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: E1202 01:36:57.082049 4884 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.083704 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.083781 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.083794 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.083811 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.083823 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:57Z","lastTransitionTime":"2025-12-02T01:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.087116 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.099358 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.110813 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.141512 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/1.log" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.147341 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.167414 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.186226 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.186303 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.186320 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.186343 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.186360 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:57Z","lastTransitionTime":"2025-12-02T01:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.187091 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-ww4r4" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.200622 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:53Z\\\",\\\"message\\\":\\\":140\\\\nI1202 01:36:53.308600 6157 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1202 01:36:53.304048 6157 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:53.308985 6157 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1202 01:36:53.309005 6157 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1202 01:36:53.308726 6157 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:53.308810 6157 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1202 01:36:53.309424 6157 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:53.309951 6157 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1202 01:36:53.310038 6157 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:53.310130 6157 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:53.310144 6157 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1202 01:36:53.310205 6157 factory.go:656] Stopping watch factory\\\\nI1202 01:36:53.310238 6157 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:53.310233 6157 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1202 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:55Z\\\",\\\"message\\\":\\\"t handler 5\\\\nI1202 01:36:55.706399 6304 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:55.706406 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1202 01:36:55.706225 6304 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706522 6304 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:55.706545 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1202 01:36:55.706553 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:55.706575 6304 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706849 6304 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:55.707271 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:55.707390 6304 factory.go:656] Stopping watch factory\\\\nI1202 01:36:55.707426 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: W1202 01:36:57.206387 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6b8c081_9419_49d2_a77f_c4fc1f136379.slice/crio-1760568cc4d318e22c0e8b98e557cf4db0cfbb1f73ae1032cf4586b8c043da55 WatchSource:0}: Error finding container 1760568cc4d318e22c0e8b98e557cf4db0cfbb1f73ae1032cf4586b8c043da55: Status 404 returned error can't find the container with id 1760568cc4d318e22c0e8b98e557cf4db0cfbb1f73ae1032cf4586b8c043da55 Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.224661 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.252963 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.270952 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.288232 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.288277 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.288295 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.288317 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.288333 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:57Z","lastTransitionTime":"2025-12-02T01:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.292698 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.309377 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.321397 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.335857 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.354641 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.369317 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.385725 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.391165 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.391222 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.391237 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.391256 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.391268 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:57Z","lastTransitionTime":"2025-12-02T01:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.398002 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.412510 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.426080 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.436892 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.454702 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.470144 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.494972 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.495010 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.495028 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.495045 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.495057 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:57Z","lastTransitionTime":"2025-12-02T01:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.495840 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs\") pod \"network-metrics-daemon-whh6t\" (UID: \"7a6a67e3-3b98-45df-9573-6efee410a90c\") " pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:36:57 crc kubenswrapper[4884]: E1202 01:36:57.495955 4884 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:36:57 crc kubenswrapper[4884]: E1202 01:36:57.496122 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs podName:7a6a67e3-3b98-45df-9573-6efee410a90c nodeName:}" failed. No retries permitted until 2025-12-02 01:36:58.496102955 +0000 UTC m=+35.171939849 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs") pod "network-metrics-daemon-whh6t" (UID: "7a6a67e3-3b98-45df-9573-6efee410a90c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.503495 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:53Z\\\",\\\"message\\\":\\\":140\\\\nI1202 01:36:53.308600 6157 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1202 01:36:53.304048 6157 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:53.308985 6157 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1202 01:36:53.309005 6157 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1202 01:36:53.308726 6157 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:53.308810 6157 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1202 01:36:53.309424 6157 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:53.309951 6157 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1202 01:36:53.310038 6157 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:53.310130 6157 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:53.310144 6157 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1202 01:36:53.310205 6157 factory.go:656] Stopping watch factory\\\\nI1202 01:36:53.310238 6157 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:53.310233 6157 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1202 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:55Z\\\",\\\"message\\\":\\\"t handler 5\\\\nI1202 01:36:55.706399 6304 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:55.706406 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1202 01:36:55.706225 6304 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706522 6304 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:55.706545 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1202 01:36:55.706553 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:55.706575 6304 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706849 6304 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:55.707271 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:55.707390 6304 factory.go:656] Stopping watch factory\\\\nI1202 01:36:55.707426 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:57Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.597477 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.597529 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.597540 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.597557 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.597571 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:57Z","lastTransitionTime":"2025-12-02T01:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.699678 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.699738 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.699780 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.699809 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.699829 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:57Z","lastTransitionTime":"2025-12-02T01:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.802202 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.802263 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.802283 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.802306 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.802326 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:57Z","lastTransitionTime":"2025-12-02T01:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.905346 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.905407 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.905426 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.905450 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:57 crc kubenswrapper[4884]: I1202 01:36:57.905469 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:57Z","lastTransitionTime":"2025-12-02T01:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.008208 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.008266 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.008282 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.008308 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.008335 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:58Z","lastTransitionTime":"2025-12-02T01:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.112052 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.112114 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.112137 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.112164 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.112179 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:58Z","lastTransitionTime":"2025-12-02T01:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.152353 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-ww4r4" event={"ID":"f6b8c081-9419-49d2-a77f-c4fc1f136379","Type":"ContainerStarted","Data":"461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5"} Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.152422 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-ww4r4" event={"ID":"f6b8c081-9419-49d2-a77f-c4fc1f136379","Type":"ContainerStarted","Data":"1760568cc4d318e22c0e8b98e557cf4db0cfbb1f73ae1032cf4586b8c043da55"} Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.170580 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.194188 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:53Z\\\",\\\"message\\\":\\\":140\\\\nI1202 01:36:53.308600 6157 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1202 01:36:53.304048 6157 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:53.308985 6157 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1202 01:36:53.309005 6157 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1202 01:36:53.308726 6157 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:53.308810 6157 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1202 01:36:53.309424 6157 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:53.309951 6157 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1202 01:36:53.310038 6157 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:53.310130 6157 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:53.310144 6157 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1202 01:36:53.310205 6157 factory.go:656] Stopping watch factory\\\\nI1202 01:36:53.310238 6157 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:53.310233 6157 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1202 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:55Z\\\",\\\"message\\\":\\\"t handler 5\\\\nI1202 01:36:55.706399 6304 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:55.706406 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1202 01:36:55.706225 6304 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706522 6304 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:55.706545 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1202 01:36:55.706553 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:55.706575 6304 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706849 6304 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:55.707271 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:55.707390 6304 factory.go:656] Stopping watch factory\\\\nI1202 01:36:55.707426 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.219606 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.219670 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.219693 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.219721 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.219744 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:58Z","lastTransitionTime":"2025-12-02T01:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.231822 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.256477 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.274516 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.292157 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.315617 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.323251 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.323332 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.323359 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.323395 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.323476 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:58Z","lastTransitionTime":"2025-12-02T01:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.341264 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.361544 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.381292 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.398508 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.404315 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.404598 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:37:14.404576259 +0000 UTC m=+51.080413173 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.414504 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.426407 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.426448 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.426465 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.426489 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.426505 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:58Z","lastTransitionTime":"2025-12-02T01:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.429510 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.450676 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.469902 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.486339 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.501161 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:36:58Z is after 2025-08-24T17:21:41Z" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.505567 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.505648 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.505706 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.505821 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.505867 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.505894 4884 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.505907 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.505943 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.505962 4884 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.505966 4884 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.505818 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.505977 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 01:37:14.505948958 +0000 UTC m=+51.181785882 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.506110 4884 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.506133 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 01:37:14.506097632 +0000 UTC m=+51.181934546 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.506184 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs\") pod \"network-metrics-daemon-whh6t\" (UID: \"7a6a67e3-3b98-45df-9573-6efee410a90c\") " pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.506248 4884 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.506261 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:37:14.506172184 +0000 UTC m=+51.182009168 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.506300 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs podName:7a6a67e3-3b98-45df-9573-6efee410a90c nodeName:}" failed. No retries permitted until 2025-12-02 01:37:00.506281186 +0000 UTC m=+37.182118100 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs") pod "network-metrics-daemon-whh6t" (UID: "7a6a67e3-3b98-45df-9573-6efee410a90c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.506338 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:37:14.506317237 +0000 UTC m=+51.182154301 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.529655 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.529707 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.529723 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.529778 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.529803 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:58Z","lastTransitionTime":"2025-12-02T01:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.613378 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.613609 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.613769 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.613794 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.614079 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.614207 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.613952 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:36:58 crc kubenswrapper[4884]: E1202 01:36:58.614381 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.632641 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.632698 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.632716 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.632736 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.632787 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:58Z","lastTransitionTime":"2025-12-02T01:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.736091 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.736151 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.736171 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.736196 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.736215 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:58Z","lastTransitionTime":"2025-12-02T01:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.839610 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.839689 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.839714 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.839833 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.839895 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:58Z","lastTransitionTime":"2025-12-02T01:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.942409 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.942492 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.942516 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.942570 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:58 crc kubenswrapper[4884]: I1202 01:36:58.942588 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:58Z","lastTransitionTime":"2025-12-02T01:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.045885 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.045943 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.045960 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.045985 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.046003 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:59Z","lastTransitionTime":"2025-12-02T01:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.148892 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.148950 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.148969 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.148992 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.149010 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:59Z","lastTransitionTime":"2025-12-02T01:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.252517 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.252577 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.252594 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.252616 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.252638 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:59Z","lastTransitionTime":"2025-12-02T01:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.356026 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.356093 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.356111 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.356135 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.356153 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:59Z","lastTransitionTime":"2025-12-02T01:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.458948 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.459011 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.459030 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.459056 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.459073 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:59Z","lastTransitionTime":"2025-12-02T01:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.561178 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.561260 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.561284 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.561312 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.561335 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:59Z","lastTransitionTime":"2025-12-02T01:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.664333 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.664387 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.664408 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.664432 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.664450 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:59Z","lastTransitionTime":"2025-12-02T01:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.767441 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.767502 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.767519 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.767543 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.767561 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:59Z","lastTransitionTime":"2025-12-02T01:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.870773 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.870837 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.870854 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.870879 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.870896 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:59Z","lastTransitionTime":"2025-12-02T01:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.973609 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.973683 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.973709 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.973780 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:36:59 crc kubenswrapper[4884]: I1202 01:36:59.973802 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:36:59Z","lastTransitionTime":"2025-12-02T01:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.076963 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.077016 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.077034 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.077059 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.077079 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:00Z","lastTransitionTime":"2025-12-02T01:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.180271 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.180383 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.180403 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.180430 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.180449 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:00Z","lastTransitionTime":"2025-12-02T01:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.283025 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.283089 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.283113 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.283140 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.283155 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:00Z","lastTransitionTime":"2025-12-02T01:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.386645 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.386716 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.386739 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.386810 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.386833 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:00Z","lastTransitionTime":"2025-12-02T01:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.489210 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.489277 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.489294 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.489318 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.489337 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:00Z","lastTransitionTime":"2025-12-02T01:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.528313 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs\") pod \"network-metrics-daemon-whh6t\" (UID: \"7a6a67e3-3b98-45df-9573-6efee410a90c\") " pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:00 crc kubenswrapper[4884]: E1202 01:37:00.528577 4884 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:37:00 crc kubenswrapper[4884]: E1202 01:37:00.528681 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs podName:7a6a67e3-3b98-45df-9573-6efee410a90c nodeName:}" failed. No retries permitted until 2025-12-02 01:37:04.528655278 +0000 UTC m=+41.204492202 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs") pod "network-metrics-daemon-whh6t" (UID: "7a6a67e3-3b98-45df-9573-6efee410a90c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.592329 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.592383 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.592399 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.592422 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.592440 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:00Z","lastTransitionTime":"2025-12-02T01:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.613284 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.613403 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.613413 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:00 crc kubenswrapper[4884]: E1202 01:37:00.613478 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.613500 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:00 crc kubenswrapper[4884]: E1202 01:37:00.613648 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:00 crc kubenswrapper[4884]: E1202 01:37:00.613814 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:00 crc kubenswrapper[4884]: E1202 01:37:00.613900 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.695499 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.695555 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.695572 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.695608 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.695625 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:00Z","lastTransitionTime":"2025-12-02T01:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.798513 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.798564 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.798581 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.798602 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.798619 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:00Z","lastTransitionTime":"2025-12-02T01:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.901479 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.901554 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.901573 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.901600 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:00 crc kubenswrapper[4884]: I1202 01:37:00.901618 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:00Z","lastTransitionTime":"2025-12-02T01:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.004552 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.004614 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.004632 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.004658 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.004675 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:01Z","lastTransitionTime":"2025-12-02T01:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.108052 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.108133 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.108154 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.108198 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.108230 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:01Z","lastTransitionTime":"2025-12-02T01:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.211311 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.211368 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.211386 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.211409 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.211426 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:01Z","lastTransitionTime":"2025-12-02T01:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.314593 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.314653 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.314671 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.314696 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.314728 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:01Z","lastTransitionTime":"2025-12-02T01:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.418201 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.418267 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.418283 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.418305 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.418324 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:01Z","lastTransitionTime":"2025-12-02T01:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.521837 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.521900 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.521922 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.521950 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.521972 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:01Z","lastTransitionTime":"2025-12-02T01:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.625622 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.625723 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.625802 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.625832 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.625850 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:01Z","lastTransitionTime":"2025-12-02T01:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.728401 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.728454 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.728473 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.728496 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.728514 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:01Z","lastTransitionTime":"2025-12-02T01:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.831250 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.831334 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.831353 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.831377 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.831394 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:01Z","lastTransitionTime":"2025-12-02T01:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.934448 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.934528 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.934549 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.934577 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:01 crc kubenswrapper[4884]: I1202 01:37:01.934595 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:01Z","lastTransitionTime":"2025-12-02T01:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.038169 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.038238 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.038255 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.038281 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.038302 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:02Z","lastTransitionTime":"2025-12-02T01:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.140834 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.140894 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.140915 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.140944 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.140965 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:02Z","lastTransitionTime":"2025-12-02T01:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.243311 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.243365 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.243382 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.243404 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.243421 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:02Z","lastTransitionTime":"2025-12-02T01:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.346399 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.346491 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.346508 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.346537 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.346562 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:02Z","lastTransitionTime":"2025-12-02T01:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.449605 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.449667 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.449684 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.449711 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.449728 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:02Z","lastTransitionTime":"2025-12-02T01:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.492949 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.514310 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.537140 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.552072 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.552125 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.552143 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.552168 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.552187 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:02Z","lastTransitionTime":"2025-12-02T01:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.555848 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.575952 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.595410 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.614040 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.614088 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.614112 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.614124 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:02 crc kubenswrapper[4884]: E1202 01:37:02.614216 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:02 crc kubenswrapper[4884]: E1202 01:37:02.614417 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:02 crc kubenswrapper[4884]: E1202 01:37:02.614523 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:02 crc kubenswrapper[4884]: E1202 01:37:02.614690 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.616387 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.634309 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.651101 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.655069 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.655117 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.655133 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.655156 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.655171 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:02Z","lastTransitionTime":"2025-12-02T01:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.665832 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.682997 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.700130 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.718848 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.733002 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.746304 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.757859 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.757921 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.757939 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.757963 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.757980 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:02Z","lastTransitionTime":"2025-12-02T01:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.763193 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.790630 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:53Z\\\",\\\"message\\\":\\\":140\\\\nI1202 01:36:53.308600 6157 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1202 01:36:53.304048 6157 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:53.308985 6157 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1202 01:36:53.309005 6157 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1202 01:36:53.308726 6157 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:53.308810 6157 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1202 01:36:53.309424 6157 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:53.309951 6157 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1202 01:36:53.310038 6157 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:53.310130 6157 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:53.310144 6157 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1202 01:36:53.310205 6157 factory.go:656] Stopping watch factory\\\\nI1202 01:36:53.310238 6157 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:53.310233 6157 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1202 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:55Z\\\",\\\"message\\\":\\\"t handler 5\\\\nI1202 01:36:55.706399 6304 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:55.706406 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1202 01:36:55.706225 6304 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706522 6304 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:55.706545 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1202 01:36:55.706553 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:55.706575 6304 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706849 6304 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:55.707271 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:55.707390 6304 factory.go:656] Stopping watch factory\\\\nI1202 01:36:55.707426 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.839656 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:02Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.860707 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.860785 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.860803 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.860827 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.860848 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:02Z","lastTransitionTime":"2025-12-02T01:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.963723 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.963838 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.963861 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.963891 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:02 crc kubenswrapper[4884]: I1202 01:37:02.963918 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:02Z","lastTransitionTime":"2025-12-02T01:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.066776 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.066848 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.066883 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.066915 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.066942 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:03Z","lastTransitionTime":"2025-12-02T01:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.171072 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.171146 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.171169 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.171195 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.171213 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:03Z","lastTransitionTime":"2025-12-02T01:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.274472 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.274531 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.274548 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.274593 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.274612 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:03Z","lastTransitionTime":"2025-12-02T01:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.378056 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.378128 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.378151 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.378180 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.378200 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:03Z","lastTransitionTime":"2025-12-02T01:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.481843 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.481891 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.481903 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.481923 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.481935 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:03Z","lastTransitionTime":"2025-12-02T01:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.584562 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.584611 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.584622 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.584640 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.584653 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:03Z","lastTransitionTime":"2025-12-02T01:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.633545 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.651215 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.674590 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.687518 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.687576 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.687592 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.687613 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.687628 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:03Z","lastTransitionTime":"2025-12-02T01:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.690919 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.702214 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.713583 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.728691 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.742357 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.761062 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.776806 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.790026 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.790084 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.790103 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.790127 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.790145 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:03Z","lastTransitionTime":"2025-12-02T01:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.792952 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.805180 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.816105 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.827579 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.847232 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.862832 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.887485 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c251bfadfc5821729b61dd4d4ef926a1b839391e5e058d4325c3e3d71a495cc\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:53Z\\\",\\\"message\\\":\\\":140\\\\nI1202 01:36:53.308600 6157 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1202 01:36:53.304048 6157 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:53.308985 6157 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1202 01:36:53.309005 6157 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1202 01:36:53.308726 6157 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:53.308810 6157 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1202 01:36:53.309424 6157 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:53.309951 6157 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1202 01:36:53.310038 6157 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:53.310130 6157 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:53.310144 6157 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1202 01:36:53.310205 6157 factory.go:656] Stopping watch factory\\\\nI1202 01:36:53.310238 6157 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:53.310233 6157 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1202 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:55Z\\\",\\\"message\\\":\\\"t handler 5\\\\nI1202 01:36:55.706399 6304 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:55.706406 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1202 01:36:55.706225 6304 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706522 6304 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:55.706545 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1202 01:36:55.706553 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:55.706575 6304 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706849 6304 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:55.707271 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:55.707390 6304 factory.go:656] Stopping watch factory\\\\nI1202 01:36:55.707426 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:03Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.892918 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.892956 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.892967 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.892983 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.892995 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:03Z","lastTransitionTime":"2025-12-02T01:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.995917 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.995960 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.995972 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.995991 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:03 crc kubenswrapper[4884]: I1202 01:37:03.996003 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:03Z","lastTransitionTime":"2025-12-02T01:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.097796 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.097855 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.097878 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.097905 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.097925 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:04Z","lastTransitionTime":"2025-12-02T01:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.201024 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.201083 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.201103 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.201127 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.201145 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:04Z","lastTransitionTime":"2025-12-02T01:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.304092 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.304157 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.304174 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.304203 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.304221 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:04Z","lastTransitionTime":"2025-12-02T01:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.406667 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.406809 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.406834 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.406863 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.406914 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:04Z","lastTransitionTime":"2025-12-02T01:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.510245 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.510286 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.510299 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.510318 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.510331 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:04Z","lastTransitionTime":"2025-12-02T01:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.574141 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs\") pod \"network-metrics-daemon-whh6t\" (UID: \"7a6a67e3-3b98-45df-9573-6efee410a90c\") " pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:04 crc kubenswrapper[4884]: E1202 01:37:04.574405 4884 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:37:04 crc kubenswrapper[4884]: E1202 01:37:04.574527 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs podName:7a6a67e3-3b98-45df-9573-6efee410a90c nodeName:}" failed. No retries permitted until 2025-12-02 01:37:12.574495557 +0000 UTC m=+49.250332521 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs") pod "network-metrics-daemon-whh6t" (UID: "7a6a67e3-3b98-45df-9573-6efee410a90c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.613492 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.613538 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.613709 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:04 crc kubenswrapper[4884]: E1202 01:37:04.613855 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:04 crc kubenswrapper[4884]: E1202 01:37:04.614017 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:04 crc kubenswrapper[4884]: E1202 01:37:04.614113 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.613527 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.614305 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:04 crc kubenswrapper[4884]: E1202 01:37:04.614337 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.614357 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.614382 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.614412 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.614435 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:04Z","lastTransitionTime":"2025-12-02T01:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.717925 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.717979 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.717996 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.718017 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.718034 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:04Z","lastTransitionTime":"2025-12-02T01:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.821556 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.821603 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.821619 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.821643 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.821661 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:04Z","lastTransitionTime":"2025-12-02T01:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.925585 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.926029 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.926215 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.926398 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:04 crc kubenswrapper[4884]: I1202 01:37:04.926619 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:04Z","lastTransitionTime":"2025-12-02T01:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.030344 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.030404 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.030422 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.030445 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.030465 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:05Z","lastTransitionTime":"2025-12-02T01:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.133858 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.133906 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.133922 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.133946 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.133962 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:05Z","lastTransitionTime":"2025-12-02T01:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.237033 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.237103 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.237120 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.237144 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.237162 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:05Z","lastTransitionTime":"2025-12-02T01:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.340183 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.340263 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.340284 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.340313 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.340335 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:05Z","lastTransitionTime":"2025-12-02T01:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.443538 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.443594 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.443611 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.443635 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.443653 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:05Z","lastTransitionTime":"2025-12-02T01:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.546966 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.547059 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.547078 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.547106 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.547125 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:05Z","lastTransitionTime":"2025-12-02T01:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.650320 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.650405 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.650430 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.650470 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.650498 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:05Z","lastTransitionTime":"2025-12-02T01:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.753287 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.753358 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.753376 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.753404 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.753421 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:05Z","lastTransitionTime":"2025-12-02T01:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.856266 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.856344 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.856365 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.856393 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.856412 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:05Z","lastTransitionTime":"2025-12-02T01:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.960299 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.960358 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.960374 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.960397 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:05 crc kubenswrapper[4884]: I1202 01:37:05.960416 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:05Z","lastTransitionTime":"2025-12-02T01:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.063913 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.063962 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.063981 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.064004 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.064021 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:06Z","lastTransitionTime":"2025-12-02T01:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.166724 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.166811 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.166833 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.166858 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.166877 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:06Z","lastTransitionTime":"2025-12-02T01:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.269789 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.270250 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.270464 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.270674 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.271140 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:06Z","lastTransitionTime":"2025-12-02T01:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.373806 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.373872 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.373895 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.373923 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.373947 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:06Z","lastTransitionTime":"2025-12-02T01:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.477151 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.477201 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.477218 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.477240 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.477257 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:06Z","lastTransitionTime":"2025-12-02T01:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.580495 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.580553 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.580571 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.580594 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.580611 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:06Z","lastTransitionTime":"2025-12-02T01:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.613321 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.613417 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.613528 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:06 crc kubenswrapper[4884]: E1202 01:37:06.613522 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.613600 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:06 crc kubenswrapper[4884]: E1202 01:37:06.613716 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:06 crc kubenswrapper[4884]: E1202 01:37:06.613778 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:06 crc kubenswrapper[4884]: E1202 01:37:06.613846 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.684660 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.684736 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.684785 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.684811 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.684829 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:06Z","lastTransitionTime":"2025-12-02T01:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.788289 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.788358 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.788381 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.788454 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.788542 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:06Z","lastTransitionTime":"2025-12-02T01:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.891908 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.891964 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.891987 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.892015 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.892040 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:06Z","lastTransitionTime":"2025-12-02T01:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.994735 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.994994 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.995013 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.995033 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:06 crc kubenswrapper[4884]: I1202 01:37:06.995050 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:06Z","lastTransitionTime":"2025-12-02T01:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.098116 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.098176 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.098193 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.098217 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.098235 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:07Z","lastTransitionTime":"2025-12-02T01:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.201833 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.202232 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.202378 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.202516 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.202648 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:07Z","lastTransitionTime":"2025-12-02T01:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.305966 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.306304 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.306453 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.306586 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.306713 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:07Z","lastTransitionTime":"2025-12-02T01:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.371935 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.372297 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.372435 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.372572 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.372695 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:07Z","lastTransitionTime":"2025-12-02T01:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:07 crc kubenswrapper[4884]: E1202 01:37:07.395654 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.400590 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.400640 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.400651 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.400667 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.400680 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:07Z","lastTransitionTime":"2025-12-02T01:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:07 crc kubenswrapper[4884]: E1202 01:37:07.416662 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.420852 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.420882 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.420893 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.420910 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.420922 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:07Z","lastTransitionTime":"2025-12-02T01:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:07 crc kubenswrapper[4884]: E1202 01:37:07.434187 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.438219 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.438256 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.438267 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.438282 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.438293 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:07Z","lastTransitionTime":"2025-12-02T01:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:07 crc kubenswrapper[4884]: E1202 01:37:07.450221 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.454109 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.454161 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.454179 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.454203 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.454221 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:07Z","lastTransitionTime":"2025-12-02T01:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:07 crc kubenswrapper[4884]: E1202 01:37:07.471573 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: E1202 01:37:07.471723 4884 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.473635 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.473903 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.473988 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.474070 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.474165 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:07Z","lastTransitionTime":"2025-12-02T01:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.576545 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.576608 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.576626 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.576654 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.576670 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:07Z","lastTransitionTime":"2025-12-02T01:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.614613 4884 scope.go:117] "RemoveContainer" containerID="f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.637584 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.657764 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.677360 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.679720 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.679799 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.679818 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.679845 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.679863 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:07Z","lastTransitionTime":"2025-12-02T01:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.699812 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.719634 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.735138 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.757357 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.774500 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.782846 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.782881 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.782890 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.782903 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.782913 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:07Z","lastTransitionTime":"2025-12-02T01:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.789389 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.804427 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.820927 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.837572 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.849828 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.859358 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.879651 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.885777 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.885808 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.885817 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.885831 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.885840 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:07Z","lastTransitionTime":"2025-12-02T01:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.891064 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.906602 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:55Z\\\",\\\"message\\\":\\\"t handler 5\\\\nI1202 01:36:55.706399 6304 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:55.706406 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1202 01:36:55.706225 6304 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706522 6304 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:55.706545 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1202 01:36:55.706553 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:55.706575 6304 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706849 6304 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:55.707271 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:55.707390 6304 factory.go:656] Stopping watch factory\\\\nI1202 01:36:55.707426 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:07Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.988403 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.988433 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.988441 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.988455 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:07 crc kubenswrapper[4884]: I1202 01:37:07.988464 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:07Z","lastTransitionTime":"2025-12-02T01:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.090765 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.090805 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.090820 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.090838 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.090850 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:08Z","lastTransitionTime":"2025-12-02T01:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.189156 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/1.log" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.191484 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerStarted","Data":"10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88"} Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.191729 4884 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.192353 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.192410 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.192427 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.192445 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.192914 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:08Z","lastTransitionTime":"2025-12-02T01:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.206855 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.221149 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.234849 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.247637 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.258651 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.268973 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.280927 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.294287 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.295742 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.295780 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.295788 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.295802 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.295820 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:08Z","lastTransitionTime":"2025-12-02T01:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.304456 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.324684 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.337196 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.357176 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:55Z\\\",\\\"message\\\":\\\"t handler 5\\\\nI1202 01:36:55.706399 6304 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:55.706406 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1202 01:36:55.706225 6304 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706522 6304 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:55.706545 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1202 01:36:55.706553 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:55.706575 6304 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706849 6304 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:55.707271 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:55.707390 6304 factory.go:656] Stopping watch factory\\\\nI1202 01:36:55.707426 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.377902 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.394009 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.397383 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.398008 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.398103 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.398224 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.398307 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:08Z","lastTransitionTime":"2025-12-02T01:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.407902 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.420877 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.436870 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:08Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.501284 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.501923 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.501949 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.501981 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.501992 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:08Z","lastTransitionTime":"2025-12-02T01:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.604664 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.604718 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.604730 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.604771 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.604784 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:08Z","lastTransitionTime":"2025-12-02T01:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.614057 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.614057 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.614187 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.614072 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:08 crc kubenswrapper[4884]: E1202 01:37:08.614310 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:08 crc kubenswrapper[4884]: E1202 01:37:08.614456 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:08 crc kubenswrapper[4884]: E1202 01:37:08.614595 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:08 crc kubenswrapper[4884]: E1202 01:37:08.614772 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.707407 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.707455 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.707465 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.707484 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.707495 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:08Z","lastTransitionTime":"2025-12-02T01:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.810401 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.810438 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.810456 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.810497 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.810508 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:08Z","lastTransitionTime":"2025-12-02T01:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.914074 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.914463 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.914687 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.914897 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:08 crc kubenswrapper[4884]: I1202 01:37:08.915047 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:08Z","lastTransitionTime":"2025-12-02T01:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.018107 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.018184 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.018243 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.018267 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.018283 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:09Z","lastTransitionTime":"2025-12-02T01:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.121321 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.121687 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.121863 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.122033 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.122166 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:09Z","lastTransitionTime":"2025-12-02T01:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.197649 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/2.log" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.198874 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/1.log" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.203959 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8c50c39-2888-444e-bf55-49b091559978" containerID="10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88" exitCode=1 Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.204019 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerDied","Data":"10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88"} Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.204069 4884 scope.go:117] "RemoveContainer" containerID="f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.205239 4884 scope.go:117] "RemoveContainer" containerID="10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88" Dec 02 01:37:09 crc kubenswrapper[4884]: E1202 01:37:09.205499 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.226421 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.226490 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.226509 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.226533 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.226552 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:09Z","lastTransitionTime":"2025-12-02T01:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.228706 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.248266 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.265617 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.300380 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.320736 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.328832 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.328901 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.328919 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.328944 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.328964 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:09Z","lastTransitionTime":"2025-12-02T01:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.353960 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:55Z\\\",\\\"message\\\":\\\"t handler 5\\\\nI1202 01:36:55.706399 6304 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:55.706406 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1202 01:36:55.706225 6304 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706522 6304 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:55.706545 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1202 01:36:55.706553 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:55.706575 6304 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706849 6304 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:55.707271 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:55.707390 6304 factory.go:656] Stopping watch factory\\\\nI1202 01:36:55.707426 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:08Z\\\",\\\"message\\\":\\\"t:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00013561f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1202 01:37:08.550660 6529 model_client.go:382] Update operations generated as: [{Op:update Table:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.378860 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.397624 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.416778 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.431388 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.431485 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.431507 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.431537 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.431553 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:09Z","lastTransitionTime":"2025-12-02T01:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.441533 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.467067 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.489397 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.509852 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.530797 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.533841 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.533908 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.533934 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.533963 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.533985 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:09Z","lastTransitionTime":"2025-12-02T01:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.548154 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.562560 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.578486 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:09Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.636603 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.636670 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.636695 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.636722 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.636799 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:09Z","lastTransitionTime":"2025-12-02T01:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.739523 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.739581 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.739597 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.739619 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.739636 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:09Z","lastTransitionTime":"2025-12-02T01:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.824272 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.842794 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.842842 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.842858 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.842878 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.842894 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:09Z","lastTransitionTime":"2025-12-02T01:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.946303 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.946359 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.946375 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.946399 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:09 crc kubenswrapper[4884]: I1202 01:37:09.946419 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:09Z","lastTransitionTime":"2025-12-02T01:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.050013 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.050070 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.050087 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.050109 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.050126 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:10Z","lastTransitionTime":"2025-12-02T01:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.075350 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.089921 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.101224 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.117693 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.141491 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.153912 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.153982 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.154008 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.154041 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.154063 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:10Z","lastTransitionTime":"2025-12-02T01:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.164383 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.180366 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.197116 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.210510 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/2.log" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.217457 4884 scope.go:117] "RemoveContainer" containerID="10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88" Dec 02 01:37:10 crc kubenswrapper[4884]: E1202 01:37:10.217730 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.218915 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.237529 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.254042 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.256595 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.256669 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.256688 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.256714 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.256803 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:10Z","lastTransitionTime":"2025-12-02T01:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.286621 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.307317 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.339081 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5b290cda076e735610f4cb6a8d7775a606f66c3985b2010da06bcc99da09e2c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:36:55Z\\\",\\\"message\\\":\\\"t handler 5\\\\nI1202 01:36:55.706399 6304 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1202 01:36:55.706406 6304 handler.go:208] Removed *v1.Node event handler 2\\\\nI1202 01:36:55.706225 6304 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706522 6304 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1202 01:36:55.706545 6304 handler.go:208] Removed *v1.Node event handler 7\\\\nI1202 01:36:55.706553 6304 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1202 01:36:55.706575 6304 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1202 01:36:55.706849 6304 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1202 01:36:55.707271 6304 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1202 01:36:55.707390 6304 factory.go:656] Stopping watch factory\\\\nI1202 01:36:55.707426 6304 ovnkube.go:599] Stopped ovnkube\\\\nI1202 01:36:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:08Z\\\",\\\"message\\\":\\\"t:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00013561f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1202 01:37:08.550660 6529 model_client.go:382] Update operations generated as: [{Op:update Table:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.360854 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.360917 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.360936 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.360960 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.360980 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:10Z","lastTransitionTime":"2025-12-02T01:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.361083 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.385833 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.410172 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.434270 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.452923 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.463501 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.463558 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.463575 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.463601 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.463620 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:10Z","lastTransitionTime":"2025-12-02T01:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.472494 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.491326 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.504845 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.538156 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.558103 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.566185 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.566241 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.566262 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.566288 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.566308 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:10Z","lastTransitionTime":"2025-12-02T01:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.580493 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:08Z\\\",\\\"message\\\":\\\"t:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00013561f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1202 01:37:08.550660 6529 model_client.go:382] Update operations generated as: [{Op:update Table:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:37:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.594320 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aafc3731-36a9-4f87-ba3a-3966b90bd016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27f06f9f8e8f076b46dd2ec9d609217fa18bc94c8f2cc7bf6a6e530820f74e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21ab31a5e5a2b1ae8b6b2d1ee1dcc1aefcc6efe942947fa71b801f90dc346f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a218e1965e88c816e5dc27a6bfe7c6043522794f68c4eca80982082c1c3b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.609319 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.613644 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.613802 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:10 crc kubenswrapper[4884]: E1202 01:37:10.613879 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.613948 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:10 crc kubenswrapper[4884]: E1202 01:37:10.614106 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:10 crc kubenswrapper[4884]: E1202 01:37:10.614204 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.614411 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:10 crc kubenswrapper[4884]: E1202 01:37:10.614688 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.630482 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.649421 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.666475 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.668611 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.668844 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.669043 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.669235 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.669396 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:10Z","lastTransitionTime":"2025-12-02T01:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.689938 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.717198 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.731905 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.749970 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.767612 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.773442 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.773467 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.773477 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.773492 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.773503 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:10Z","lastTransitionTime":"2025-12-02T01:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.783831 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.799867 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:10Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.876294 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.876586 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.876710 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.876875 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.877002 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:10Z","lastTransitionTime":"2025-12-02T01:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.979687 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.979803 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.979832 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.979859 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:10 crc kubenswrapper[4884]: I1202 01:37:10.979877 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:10Z","lastTransitionTime":"2025-12-02T01:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.082736 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.082806 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.082819 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.082839 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.082850 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:11Z","lastTransitionTime":"2025-12-02T01:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.185655 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.186102 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.186297 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.186487 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.186670 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:11Z","lastTransitionTime":"2025-12-02T01:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.290356 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.290413 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.290431 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.290501 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.290521 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:11Z","lastTransitionTime":"2025-12-02T01:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.393605 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.393665 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.393683 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.393711 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.393729 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:11Z","lastTransitionTime":"2025-12-02T01:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.496399 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.497044 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.497079 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.497098 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.497111 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:11Z","lastTransitionTime":"2025-12-02T01:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.600325 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.600375 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.600392 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.600419 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.600439 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:11Z","lastTransitionTime":"2025-12-02T01:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.703407 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.703477 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.703496 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.703518 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.703534 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:11Z","lastTransitionTime":"2025-12-02T01:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.806987 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.807102 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.807122 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.807144 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.807155 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:11Z","lastTransitionTime":"2025-12-02T01:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.910397 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.910729 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.910911 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.911079 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:11 crc kubenswrapper[4884]: I1202 01:37:11.911211 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:11Z","lastTransitionTime":"2025-12-02T01:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.013636 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.013693 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.013713 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.013737 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.013786 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:12Z","lastTransitionTime":"2025-12-02T01:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.116282 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.116320 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.116331 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.116347 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.116357 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:12Z","lastTransitionTime":"2025-12-02T01:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.219780 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.219822 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.219832 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.219845 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.219854 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:12Z","lastTransitionTime":"2025-12-02T01:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.322351 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.322394 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.322405 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.322421 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.322433 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:12Z","lastTransitionTime":"2025-12-02T01:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.425388 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.425451 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.425468 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.425492 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.425512 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:12Z","lastTransitionTime":"2025-12-02T01:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.528604 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.528681 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.528703 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.528732 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.528784 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:12Z","lastTransitionTime":"2025-12-02T01:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.614010 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.614093 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.614040 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.614010 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:12 crc kubenswrapper[4884]: E1202 01:37:12.614235 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:12 crc kubenswrapper[4884]: E1202 01:37:12.614341 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:12 crc kubenswrapper[4884]: E1202 01:37:12.614474 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:12 crc kubenswrapper[4884]: E1202 01:37:12.614660 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.631412 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.631465 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.631475 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.631488 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.631516 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:12Z","lastTransitionTime":"2025-12-02T01:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.667852 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs\") pod \"network-metrics-daemon-whh6t\" (UID: \"7a6a67e3-3b98-45df-9573-6efee410a90c\") " pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:12 crc kubenswrapper[4884]: E1202 01:37:12.668089 4884 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:37:12 crc kubenswrapper[4884]: E1202 01:37:12.668189 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs podName:7a6a67e3-3b98-45df-9573-6efee410a90c nodeName:}" failed. No retries permitted until 2025-12-02 01:37:28.668163035 +0000 UTC m=+65.343999959 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs") pod "network-metrics-daemon-whh6t" (UID: "7a6a67e3-3b98-45df-9573-6efee410a90c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.734357 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.734433 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.734467 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.734497 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.734518 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:12Z","lastTransitionTime":"2025-12-02T01:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.837309 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.837363 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.837374 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.837392 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.837405 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:12Z","lastTransitionTime":"2025-12-02T01:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.940259 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.940325 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.940343 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.940367 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:12 crc kubenswrapper[4884]: I1202 01:37:12.940388 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:12Z","lastTransitionTime":"2025-12-02T01:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.043168 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.043204 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.043215 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.043231 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.043245 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:13Z","lastTransitionTime":"2025-12-02T01:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.151051 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.151285 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.151301 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.151325 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.151342 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:13Z","lastTransitionTime":"2025-12-02T01:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.253919 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.253987 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.254009 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.254038 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.254058 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:13Z","lastTransitionTime":"2025-12-02T01:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.356663 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.356737 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.356789 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.356821 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.356844 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:13Z","lastTransitionTime":"2025-12-02T01:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.459529 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.459588 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.459611 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.459638 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.459658 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:13Z","lastTransitionTime":"2025-12-02T01:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.563810 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.563920 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.563938 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.564001 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.564018 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:13Z","lastTransitionTime":"2025-12-02T01:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.647613 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.667310 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.667431 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.667454 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.667531 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.667547 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:13Z","lastTransitionTime":"2025-12-02T01:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.671827 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.706119 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:08Z\\\",\\\"message\\\":\\\"t:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00013561f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1202 01:37:08.550660 6529 model_client.go:382] Update operations generated as: [{Op:update Table:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:37:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.731150 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.756084 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.771020 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.771080 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.771097 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.771124 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.771145 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:13Z","lastTransitionTime":"2025-12-02T01:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.776065 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aafc3731-36a9-4f87-ba3a-3966b90bd016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27f06f9f8e8f076b46dd2ec9d609217fa18bc94c8f2cc7bf6a6e530820f74e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21ab31a5e5a2b1ae8b6b2d1ee1dcc1aefcc6efe942947fa71b801f90dc346f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a218e1965e88c816e5dc27a6bfe7c6043522794f68c4eca80982082c1c3b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.797227 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.816220 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.834738 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.852250 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.870855 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.874216 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.874274 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.874292 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.874316 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.874335 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:13Z","lastTransitionTime":"2025-12-02T01:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.897079 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.917909 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.937454 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.969240 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.976731 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.976786 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.976798 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.976814 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.976827 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:13Z","lastTransitionTime":"2025-12-02T01:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:13 crc kubenswrapper[4884]: I1202 01:37:13.996346 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:13Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.016768 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:14Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.028951 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:14Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.080223 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.080286 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.080307 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.080331 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.080350 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:14Z","lastTransitionTime":"2025-12-02T01:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.183650 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.184125 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.184144 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.184175 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.184201 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:14Z","lastTransitionTime":"2025-12-02T01:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.287513 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.288068 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.288350 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.288527 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.288681 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:14Z","lastTransitionTime":"2025-12-02T01:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.392214 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.392289 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.392309 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.392337 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.392360 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:14Z","lastTransitionTime":"2025-12-02T01:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.486443 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.486676 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:37:46.486643991 +0000 UTC m=+83.162480915 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.495052 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.495264 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.495636 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.496037 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.496379 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:14Z","lastTransitionTime":"2025-12-02T01:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.587648 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.587711 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.587773 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.587823 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.588027 4884 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.588121 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:37:46.588097952 +0000 UTC m=+83.263934866 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.588335 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.588361 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.588379 4884 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.588423 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 01:37:46.58840917 +0000 UTC m=+83.264246084 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.588462 4884 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.588503 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.588526 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.588540 4884 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.588587 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 01:37:46.588574904 +0000 UTC m=+83.264411828 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.588610 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:37:46.588598614 +0000 UTC m=+83.264435528 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.600925 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.601150 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.601285 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.601418 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.601568 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:14Z","lastTransitionTime":"2025-12-02T01:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.613347 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.613409 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.613481 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.613920 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.614042 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.613509 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.614178 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:14 crc kubenswrapper[4884]: E1202 01:37:14.613699 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.705091 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.705154 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.705171 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.705197 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.705220 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:14Z","lastTransitionTime":"2025-12-02T01:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.807864 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.807940 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.807957 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.807985 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.808005 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:14Z","lastTransitionTime":"2025-12-02T01:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.911103 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.911136 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.911177 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.911195 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:14 crc kubenswrapper[4884]: I1202 01:37:14.911207 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:14Z","lastTransitionTime":"2025-12-02T01:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.014603 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.014954 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.015152 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.015386 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.015591 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:15Z","lastTransitionTime":"2025-12-02T01:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.118541 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.118606 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.118631 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.118658 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.118679 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:15Z","lastTransitionTime":"2025-12-02T01:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.221574 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.221633 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.221656 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.221685 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.221706 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:15Z","lastTransitionTime":"2025-12-02T01:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.324343 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.324405 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.324427 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.324455 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.324478 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:15Z","lastTransitionTime":"2025-12-02T01:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.427542 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.427606 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.427630 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.427657 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.427678 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:15Z","lastTransitionTime":"2025-12-02T01:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.530639 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.530713 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.530737 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.530827 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.530853 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:15Z","lastTransitionTime":"2025-12-02T01:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.633993 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.634059 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.634080 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.634105 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.634127 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:15Z","lastTransitionTime":"2025-12-02T01:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.736713 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.736828 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.736857 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.736881 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.736898 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:15Z","lastTransitionTime":"2025-12-02T01:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.839695 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.839797 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.839823 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.839846 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.839864 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:15Z","lastTransitionTime":"2025-12-02T01:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.942823 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.942882 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.942898 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.942920 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:15 crc kubenswrapper[4884]: I1202 01:37:15.942940 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:15Z","lastTransitionTime":"2025-12-02T01:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.050672 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.050770 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.050797 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.050828 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.050849 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:16Z","lastTransitionTime":"2025-12-02T01:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.153959 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.154047 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.154072 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.154105 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.154128 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:16Z","lastTransitionTime":"2025-12-02T01:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.260048 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.260122 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.260145 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.260174 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.260195 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:16Z","lastTransitionTime":"2025-12-02T01:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.363180 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.363266 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.363290 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.363319 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.363343 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:16Z","lastTransitionTime":"2025-12-02T01:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.471977 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.472035 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.472054 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.472080 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.472097 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:16Z","lastTransitionTime":"2025-12-02T01:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.575179 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.575238 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.575279 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.575305 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.575323 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:16Z","lastTransitionTime":"2025-12-02T01:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.614010 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.614111 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:16 crc kubenswrapper[4884]: E1202 01:37:16.614210 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.614111 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:16 crc kubenswrapper[4884]: E1202 01:37:16.614343 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.614436 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:16 crc kubenswrapper[4884]: E1202 01:37:16.614467 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:16 crc kubenswrapper[4884]: E1202 01:37:16.614723 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.679093 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.679186 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.679210 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.679247 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.679273 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:16Z","lastTransitionTime":"2025-12-02T01:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.782924 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.783040 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.783077 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.783122 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.783161 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:16Z","lastTransitionTime":"2025-12-02T01:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.886598 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.886674 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.886693 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.886716 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.886736 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:16Z","lastTransitionTime":"2025-12-02T01:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.989903 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.989964 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.989980 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.990003 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:16 crc kubenswrapper[4884]: I1202 01:37:16.990021 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:16Z","lastTransitionTime":"2025-12-02T01:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.093710 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.093832 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.093848 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.093870 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.093883 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:17Z","lastTransitionTime":"2025-12-02T01:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.196741 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.196844 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.196866 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.196896 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.196919 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:17Z","lastTransitionTime":"2025-12-02T01:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.300218 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.300279 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.300292 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.300315 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.300332 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:17Z","lastTransitionTime":"2025-12-02T01:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.403181 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.403242 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.403255 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.403278 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.403296 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:17Z","lastTransitionTime":"2025-12-02T01:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.506167 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.506229 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.506245 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.506265 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.506277 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:17Z","lastTransitionTime":"2025-12-02T01:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.555130 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.555191 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.555211 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.555236 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.555253 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:17Z","lastTransitionTime":"2025-12-02T01:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:17 crc kubenswrapper[4884]: E1202 01:37:17.577932 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:17Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.583340 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.583408 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.583427 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.583453 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.583472 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:17Z","lastTransitionTime":"2025-12-02T01:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:17 crc kubenswrapper[4884]: E1202 01:37:17.604998 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:17Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.610170 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.610288 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.610307 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.610327 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.610343 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:17Z","lastTransitionTime":"2025-12-02T01:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:17 crc kubenswrapper[4884]: E1202 01:37:17.632056 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:17Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.637442 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.637689 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.638054 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.638394 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.638733 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:17Z","lastTransitionTime":"2025-12-02T01:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:17 crc kubenswrapper[4884]: E1202 01:37:17.658502 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:17Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.663874 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.663955 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.663974 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.663999 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.664020 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:17Z","lastTransitionTime":"2025-12-02T01:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:17 crc kubenswrapper[4884]: E1202 01:37:17.685991 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:17Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:17 crc kubenswrapper[4884]: E1202 01:37:17.686217 4884 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.688528 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.688576 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.688593 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.688619 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.688637 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:17Z","lastTransitionTime":"2025-12-02T01:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.791806 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.791881 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.791906 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.791945 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.791963 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:17Z","lastTransitionTime":"2025-12-02T01:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.894948 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.895288 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.895466 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.895611 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.895727 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:17Z","lastTransitionTime":"2025-12-02T01:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.998884 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.998942 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.998959 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.998984 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:17 crc kubenswrapper[4884]: I1202 01:37:17.999002 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:17Z","lastTransitionTime":"2025-12-02T01:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.101685 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.101791 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.101836 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.101861 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.101878 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:18Z","lastTransitionTime":"2025-12-02T01:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.206048 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.206108 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.206126 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.206209 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.206239 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:18Z","lastTransitionTime":"2025-12-02T01:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.309330 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.309399 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.309418 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.309444 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.309462 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:18Z","lastTransitionTime":"2025-12-02T01:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.411929 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.412045 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.412068 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.412098 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.412120 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:18Z","lastTransitionTime":"2025-12-02T01:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.515361 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.515419 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.515436 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.515459 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.515477 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:18Z","lastTransitionTime":"2025-12-02T01:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.613139 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.613189 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.613174 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.613174 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:18 crc kubenswrapper[4884]: E1202 01:37:18.613374 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:18 crc kubenswrapper[4884]: E1202 01:37:18.613495 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:18 crc kubenswrapper[4884]: E1202 01:37:18.613599 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:18 crc kubenswrapper[4884]: E1202 01:37:18.613712 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.619122 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.619180 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.619199 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.619222 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.619240 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:18Z","lastTransitionTime":"2025-12-02T01:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.721973 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.722037 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.722059 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.722083 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.722105 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:18Z","lastTransitionTime":"2025-12-02T01:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.825002 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.825064 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.825082 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.825107 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.825124 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:18Z","lastTransitionTime":"2025-12-02T01:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.927878 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.927943 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.927961 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.927988 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:18 crc kubenswrapper[4884]: I1202 01:37:18.928005 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:18Z","lastTransitionTime":"2025-12-02T01:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.031101 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.031153 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.031171 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.031193 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.031211 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:19Z","lastTransitionTime":"2025-12-02T01:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.134037 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.134112 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.134129 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.134154 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.134172 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:19Z","lastTransitionTime":"2025-12-02T01:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.237285 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.237343 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.237360 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.237384 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.237402 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:19Z","lastTransitionTime":"2025-12-02T01:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.340078 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.340142 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.340163 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.340191 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.340211 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:19Z","lastTransitionTime":"2025-12-02T01:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.443091 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.443150 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.443170 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.443194 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.443211 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:19Z","lastTransitionTime":"2025-12-02T01:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.546601 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.546650 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.546666 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.546688 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.546705 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:19Z","lastTransitionTime":"2025-12-02T01:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.649897 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.650192 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.650226 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.650261 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.650286 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:19Z","lastTransitionTime":"2025-12-02T01:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.752959 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.753023 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.753042 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.753064 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.753083 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:19Z","lastTransitionTime":"2025-12-02T01:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.856270 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.856338 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.856354 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.856378 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.856396 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:19Z","lastTransitionTime":"2025-12-02T01:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.958902 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.958984 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.959001 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.959025 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:19 crc kubenswrapper[4884]: I1202 01:37:19.959042 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:19Z","lastTransitionTime":"2025-12-02T01:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.061837 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.061893 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.061909 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.061934 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.061951 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:20Z","lastTransitionTime":"2025-12-02T01:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.165244 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.165333 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.165358 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.165413 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.165432 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:20Z","lastTransitionTime":"2025-12-02T01:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.268410 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.268475 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.268493 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.268517 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.268537 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:20Z","lastTransitionTime":"2025-12-02T01:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.371154 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.371217 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.371250 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.371290 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.371312 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:20Z","lastTransitionTime":"2025-12-02T01:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.473593 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.473670 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.473696 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.473721 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.473773 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:20Z","lastTransitionTime":"2025-12-02T01:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.577223 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.577292 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.577309 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.577333 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.577350 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:20Z","lastTransitionTime":"2025-12-02T01:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.613494 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.613576 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.613626 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:20 crc kubenswrapper[4884]: E1202 01:37:20.613836 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.613906 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:20 crc kubenswrapper[4884]: E1202 01:37:20.613968 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:20 crc kubenswrapper[4884]: E1202 01:37:20.614115 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:20 crc kubenswrapper[4884]: E1202 01:37:20.614276 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.680028 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.680100 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.680124 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.680147 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.680167 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:20Z","lastTransitionTime":"2025-12-02T01:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.783616 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.783693 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.783716 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.783840 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.783869 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:20Z","lastTransitionTime":"2025-12-02T01:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.886612 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.886680 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.886697 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.886725 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.886779 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:20Z","lastTransitionTime":"2025-12-02T01:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.989119 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.989184 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.989206 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.989234 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:20 crc kubenswrapper[4884]: I1202 01:37:20.989258 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:20Z","lastTransitionTime":"2025-12-02T01:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.092609 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.092675 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.092698 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.092725 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.092779 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:21Z","lastTransitionTime":"2025-12-02T01:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.195982 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.196041 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.196059 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.196082 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.196103 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:21Z","lastTransitionTime":"2025-12-02T01:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.299049 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.299078 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.299087 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.299099 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.299109 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:21Z","lastTransitionTime":"2025-12-02T01:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.402124 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.402194 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.402216 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.402245 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.402262 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:21Z","lastTransitionTime":"2025-12-02T01:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.504717 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.504798 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.504816 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.504838 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.504854 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:21Z","lastTransitionTime":"2025-12-02T01:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.608208 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.608271 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.608316 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.608339 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.608357 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:21Z","lastTransitionTime":"2025-12-02T01:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.614898 4884 scope.go:117] "RemoveContainer" containerID="10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88" Dec 02 01:37:21 crc kubenswrapper[4884]: E1202 01:37:21.615151 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.710879 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.710959 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.710983 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.711014 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.711035 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:21Z","lastTransitionTime":"2025-12-02T01:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.814559 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.814602 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.814615 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.814632 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.814644 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:21Z","lastTransitionTime":"2025-12-02T01:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.917494 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.917530 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.917543 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.917561 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:21 crc kubenswrapper[4884]: I1202 01:37:21.917575 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:21Z","lastTransitionTime":"2025-12-02T01:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.021202 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.021260 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.021277 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.021303 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.021321 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:22Z","lastTransitionTime":"2025-12-02T01:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.124860 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.124922 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.124933 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.124960 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.124970 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:22Z","lastTransitionTime":"2025-12-02T01:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.227951 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.228025 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.228047 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.228072 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.228090 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:22Z","lastTransitionTime":"2025-12-02T01:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.330431 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.330508 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.330526 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.330551 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.330569 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:22Z","lastTransitionTime":"2025-12-02T01:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.433355 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.433421 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.433438 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.433470 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.433494 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:22Z","lastTransitionTime":"2025-12-02T01:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.536970 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.537039 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.537057 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.537082 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.537100 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:22Z","lastTransitionTime":"2025-12-02T01:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.613226 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.613260 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.613333 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.613230 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:22 crc kubenswrapper[4884]: E1202 01:37:22.613393 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:22 crc kubenswrapper[4884]: E1202 01:37:22.613500 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:22 crc kubenswrapper[4884]: E1202 01:37:22.614024 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:22 crc kubenswrapper[4884]: E1202 01:37:22.614124 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.640678 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.640782 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.640801 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.640838 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.640857 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:22Z","lastTransitionTime":"2025-12-02T01:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.744959 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.745021 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.745038 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.745064 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.745086 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:22Z","lastTransitionTime":"2025-12-02T01:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.848868 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.848957 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.848984 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.849016 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.849039 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:22Z","lastTransitionTime":"2025-12-02T01:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.952638 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.952700 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.952719 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.952768 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:22 crc kubenswrapper[4884]: I1202 01:37:22.952787 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:22Z","lastTransitionTime":"2025-12-02T01:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.061881 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.062064 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.062086 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.062133 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.062150 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:23Z","lastTransitionTime":"2025-12-02T01:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.165806 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.165886 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.165906 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.165934 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.165955 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:23Z","lastTransitionTime":"2025-12-02T01:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.269820 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.269894 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.269913 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.269941 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.269961 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:23Z","lastTransitionTime":"2025-12-02T01:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.374357 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.374426 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.374442 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.374470 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.374489 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:23Z","lastTransitionTime":"2025-12-02T01:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.518923 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.518990 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.519006 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.519030 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.519046 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:23Z","lastTransitionTime":"2025-12-02T01:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.622158 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.622220 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.622237 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.622261 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.622278 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:23Z","lastTransitionTime":"2025-12-02T01:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.653052 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.675605 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.710629 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:08Z\\\",\\\"message\\\":\\\"t:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00013561f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1202 01:37:08.550660 6529 model_client.go:382] Update operations generated as: [{Op:update Table:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:37:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.725657 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.725716 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.725734 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.725787 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.725806 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:23Z","lastTransitionTime":"2025-12-02T01:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.732575 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aafc3731-36a9-4f87-ba3a-3966b90bd016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27f06f9f8e8f076b46dd2ec9d609217fa18bc94c8f2cc7bf6a6e530820f74e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21ab31a5e5a2b1ae8b6b2d1ee1dcc1aefcc6efe942947fa71b801f90dc346f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a218e1965e88c816e5dc27a6bfe7c6043522794f68c4eca80982082c1c3b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.760999 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.787076 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.802904 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.823172 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.829410 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.829484 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.829503 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.829535 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.829556 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:23Z","lastTransitionTime":"2025-12-02T01:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.846855 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.869725 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.889883 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.910787 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.929147 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.933075 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.933170 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.933262 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.933292 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.933309 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:23Z","lastTransitionTime":"2025-12-02T01:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.946872 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.964100 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:23 crc kubenswrapper[4884]: I1202 01:37:23.986128 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:23Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.004939 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:24Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.021318 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:24Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.037117 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.037181 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.037200 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.037227 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.037245 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:24Z","lastTransitionTime":"2025-12-02T01:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.140295 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.140350 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.140367 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.140390 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.140407 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:24Z","lastTransitionTime":"2025-12-02T01:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.244605 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.244649 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.244662 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.244678 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.244689 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:24Z","lastTransitionTime":"2025-12-02T01:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.348279 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.348369 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.348394 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.348428 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.348453 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:24Z","lastTransitionTime":"2025-12-02T01:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.451317 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.451382 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.451400 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.451425 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.451441 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:24Z","lastTransitionTime":"2025-12-02T01:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.554870 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.554932 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.554949 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.554978 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.554995 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:24Z","lastTransitionTime":"2025-12-02T01:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.614180 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:24 crc kubenswrapper[4884]: E1202 01:37:24.614372 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.614990 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.615044 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.615084 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:24 crc kubenswrapper[4884]: E1202 01:37:24.615176 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:24 crc kubenswrapper[4884]: E1202 01:37:24.615304 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:24 crc kubenswrapper[4884]: E1202 01:37:24.615387 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.658238 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.658293 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.658312 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.658337 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.658353 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:24Z","lastTransitionTime":"2025-12-02T01:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.761548 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.761611 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.761633 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.761657 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.761675 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:24Z","lastTransitionTime":"2025-12-02T01:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.864293 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.864352 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.864373 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.864397 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.864416 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:24Z","lastTransitionTime":"2025-12-02T01:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.968530 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.968590 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.968607 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.968631 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:24 crc kubenswrapper[4884]: I1202 01:37:24.968650 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:24Z","lastTransitionTime":"2025-12-02T01:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.071858 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.071951 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.071969 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.072025 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.072052 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:25Z","lastTransitionTime":"2025-12-02T01:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.175163 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.175213 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.175229 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.175251 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.175268 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:25Z","lastTransitionTime":"2025-12-02T01:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.277681 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.277810 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.277841 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.277872 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.277895 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:25Z","lastTransitionTime":"2025-12-02T01:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.380693 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.380777 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.380795 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.380821 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.380838 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:25Z","lastTransitionTime":"2025-12-02T01:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.483660 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.483709 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.483726 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.483777 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.483794 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:25Z","lastTransitionTime":"2025-12-02T01:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.586512 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.586543 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.586550 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.586563 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.586572 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:25Z","lastTransitionTime":"2025-12-02T01:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.689866 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.689950 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.689977 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.690042 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.690072 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:25Z","lastTransitionTime":"2025-12-02T01:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.795181 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.795243 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.795260 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.795285 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.795302 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:25Z","lastTransitionTime":"2025-12-02T01:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.897434 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.897482 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.897497 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.897519 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:25 crc kubenswrapper[4884]: I1202 01:37:25.897536 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:25Z","lastTransitionTime":"2025-12-02T01:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.000483 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.000561 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.000580 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.000603 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.000620 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:26Z","lastTransitionTime":"2025-12-02T01:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.102966 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.103022 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.103039 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.103060 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.103076 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:26Z","lastTransitionTime":"2025-12-02T01:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.205402 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.205451 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.205467 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.205487 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.205502 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:26Z","lastTransitionTime":"2025-12-02T01:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.307726 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.307808 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.307826 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.307853 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.307871 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:26Z","lastTransitionTime":"2025-12-02T01:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.410445 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.410490 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.410505 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.410525 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.410544 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:26Z","lastTransitionTime":"2025-12-02T01:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.514047 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.514096 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.514108 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.514125 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.514137 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:26Z","lastTransitionTime":"2025-12-02T01:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.613241 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.613317 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.613322 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.613258 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:26 crc kubenswrapper[4884]: E1202 01:37:26.613534 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:26 crc kubenswrapper[4884]: E1202 01:37:26.613666 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:26 crc kubenswrapper[4884]: E1202 01:37:26.613819 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:26 crc kubenswrapper[4884]: E1202 01:37:26.613924 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.616366 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.616424 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.616448 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.616477 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.616506 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:26Z","lastTransitionTime":"2025-12-02T01:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.719072 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.719115 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.719127 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.719145 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.719158 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:26Z","lastTransitionTime":"2025-12-02T01:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.822483 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.822540 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.822555 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.822612 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.822627 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:26Z","lastTransitionTime":"2025-12-02T01:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.924964 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.925006 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.925019 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.925037 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:26 crc kubenswrapper[4884]: I1202 01:37:26.925048 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:26Z","lastTransitionTime":"2025-12-02T01:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.027369 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.027426 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.027438 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.027453 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.027487 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:27Z","lastTransitionTime":"2025-12-02T01:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.130353 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.130417 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.130433 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.130449 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.130461 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:27Z","lastTransitionTime":"2025-12-02T01:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.232924 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.233304 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.233527 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.233705 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.233895 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:27Z","lastTransitionTime":"2025-12-02T01:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.336676 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.336950 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.337170 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.337392 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.337602 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:27Z","lastTransitionTime":"2025-12-02T01:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.440235 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.440307 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.440318 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.440337 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.440348 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:27Z","lastTransitionTime":"2025-12-02T01:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.543486 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.543757 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.543866 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.543995 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.544091 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:27Z","lastTransitionTime":"2025-12-02T01:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.646205 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.646260 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.646277 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.646299 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.646315 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:27Z","lastTransitionTime":"2025-12-02T01:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.749468 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.750187 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.750498 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.750695 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.751069 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:27Z","lastTransitionTime":"2025-12-02T01:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.806233 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.806270 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.806280 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.806297 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.806307 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:27Z","lastTransitionTime":"2025-12-02T01:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:27 crc kubenswrapper[4884]: E1202 01:37:27.824085 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:27Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.828476 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.828538 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.828561 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.828585 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.828606 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:27Z","lastTransitionTime":"2025-12-02T01:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:27 crc kubenswrapper[4884]: E1202 01:37:27.842543 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:27Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.846630 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.846665 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.846676 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.846691 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.846719 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:27Z","lastTransitionTime":"2025-12-02T01:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:27 crc kubenswrapper[4884]: E1202 01:37:27.863108 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:27Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.867612 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.867644 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.867655 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.867669 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.867680 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:27Z","lastTransitionTime":"2025-12-02T01:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:27 crc kubenswrapper[4884]: E1202 01:37:27.886530 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:27Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.891181 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.891217 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.891228 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.891246 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.891259 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:27Z","lastTransitionTime":"2025-12-02T01:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:27 crc kubenswrapper[4884]: E1202 01:37:27.910767 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:27Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:27 crc kubenswrapper[4884]: E1202 01:37:27.910996 4884 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.913001 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.913049 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.913066 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.913086 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:27 crc kubenswrapper[4884]: I1202 01:37:27.913103 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:27Z","lastTransitionTime":"2025-12-02T01:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.015928 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.015962 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.015972 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.015985 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.015996 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:28Z","lastTransitionTime":"2025-12-02T01:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.118898 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.118962 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.118979 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.119006 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.119024 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:28Z","lastTransitionTime":"2025-12-02T01:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.221642 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.221690 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.221704 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.221721 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.221733 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:28Z","lastTransitionTime":"2025-12-02T01:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.323784 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.323814 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.323825 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.323840 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.323852 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:28Z","lastTransitionTime":"2025-12-02T01:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.426001 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.426061 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.426079 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.426104 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.426122 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:28Z","lastTransitionTime":"2025-12-02T01:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.569539 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.569583 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.569594 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.569611 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.569623 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:28Z","lastTransitionTime":"2025-12-02T01:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.613150 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.613199 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:28 crc kubenswrapper[4884]: E1202 01:37:28.613277 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.613335 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.613387 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:28 crc kubenswrapper[4884]: E1202 01:37:28.613479 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:28 crc kubenswrapper[4884]: E1202 01:37:28.613427 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:28 crc kubenswrapper[4884]: E1202 01:37:28.613616 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.671639 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.671671 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.671681 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.671696 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.671706 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:28Z","lastTransitionTime":"2025-12-02T01:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.691226 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs\") pod \"network-metrics-daemon-whh6t\" (UID: \"7a6a67e3-3b98-45df-9573-6efee410a90c\") " pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:28 crc kubenswrapper[4884]: E1202 01:37:28.691347 4884 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:37:28 crc kubenswrapper[4884]: E1202 01:37:28.691391 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs podName:7a6a67e3-3b98-45df-9573-6efee410a90c nodeName:}" failed. No retries permitted until 2025-12-02 01:38:00.691377317 +0000 UTC m=+97.367214201 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs") pod "network-metrics-daemon-whh6t" (UID: "7a6a67e3-3b98-45df-9573-6efee410a90c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.774413 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.774449 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.774460 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.774475 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.774485 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:28Z","lastTransitionTime":"2025-12-02T01:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.877711 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.877765 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.877775 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.877790 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.877802 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:28Z","lastTransitionTime":"2025-12-02T01:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.980842 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.980892 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.980908 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.980930 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:28 crc kubenswrapper[4884]: I1202 01:37:28.980948 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:28Z","lastTransitionTime":"2025-12-02T01:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.087844 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.087906 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.087924 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.087947 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.087997 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:29Z","lastTransitionTime":"2025-12-02T01:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.190201 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.190237 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.190252 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.190272 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.190288 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:29Z","lastTransitionTime":"2025-12-02T01:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.292447 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.292505 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.292522 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.292549 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.292567 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:29Z","lastTransitionTime":"2025-12-02T01:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.396824 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.397122 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.397138 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.397161 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.397179 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:29Z","lastTransitionTime":"2025-12-02T01:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.499188 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.499229 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.499240 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.499258 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.499271 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:29Z","lastTransitionTime":"2025-12-02T01:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.602652 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.602707 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.602723 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.602769 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.602786 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:29Z","lastTransitionTime":"2025-12-02T01:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.705883 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.705941 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.705961 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.705986 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.706005 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:29Z","lastTransitionTime":"2025-12-02T01:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.808781 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.808850 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.808872 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.808901 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.808923 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:29Z","lastTransitionTime":"2025-12-02T01:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.911827 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.911882 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.911895 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.911916 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:29 crc kubenswrapper[4884]: I1202 01:37:29.911928 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:29Z","lastTransitionTime":"2025-12-02T01:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.014467 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.014518 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.014535 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.014559 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.014575 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:30Z","lastTransitionTime":"2025-12-02T01:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.116886 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.116918 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.116930 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.116943 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.116954 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:30Z","lastTransitionTime":"2025-12-02T01:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.219415 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.219446 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.219457 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.219471 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.219481 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:30Z","lastTransitionTime":"2025-12-02T01:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.286813 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6l592_791af0d6-a201-430f-b09d-02c24a4b7503/kube-multus/0.log" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.286894 4884 generic.go:334] "Generic (PLEG): container finished" podID="791af0d6-a201-430f-b09d-02c24a4b7503" containerID="4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95" exitCode=1 Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.286940 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6l592" event={"ID":"791af0d6-a201-430f-b09d-02c24a4b7503","Type":"ContainerDied","Data":"4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95"} Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.287546 4884 scope.go:117] "RemoveContainer" containerID="4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.304967 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.321232 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.321902 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.321929 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.321939 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.321955 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.321967 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:30Z","lastTransitionTime":"2025-12-02T01:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.334198 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.346523 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.362121 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.383225 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.400467 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.412773 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.423430 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.423477 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.423486 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.423504 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.423513 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:30Z","lastTransitionTime":"2025-12-02T01:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.429917 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.457816 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:08Z\\\",\\\"message\\\":\\\"t:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00013561f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1202 01:37:08.550660 6529 model_client.go:382] Update operations generated as: [{Op:update Table:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:37:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.488730 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.504063 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.517729 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.526432 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.526527 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.526550 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.526580 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.526608 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:30Z","lastTransitionTime":"2025-12-02T01:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.536624 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.555114 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:30Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:29Z\\\",\\\"message\\\":\\\"2025-12-02T01:36:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61daf367-6f4b-4e21-81a0-cb44a5376679\\\\n2025-12-02T01:36:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61daf367-6f4b-4e21-81a0-cb44a5376679 to /host/opt/cni/bin/\\\\n2025-12-02T01:36:44Z [verbose] multus-daemon started\\\\n2025-12-02T01:36:44Z [verbose] Readiness Indicator file check\\\\n2025-12-02T01:37:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.577029 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.597318 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aafc3731-36a9-4f87-ba3a-3966b90bd016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27f06f9f8e8f076b46dd2ec9d609217fa18bc94c8f2cc7bf6a6e530820f74e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21ab31a5e5a2b1ae8b6b2d1ee1dcc1aefcc6efe942947fa71b801f90dc346f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a218e1965e88c816e5dc27a6bfe7c6043522794f68c4eca80982082c1c3b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.613272 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.613370 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.613299 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:30 crc kubenswrapper[4884]: E1202 01:37:30.613470 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.613506 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:30 crc kubenswrapper[4884]: E1202 01:37:30.613631 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:30 crc kubenswrapper[4884]: E1202 01:37:30.613766 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:30 crc kubenswrapper[4884]: E1202 01:37:30.613884 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.616917 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:30Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.629099 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.629140 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.629156 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.629180 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.629197 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:30Z","lastTransitionTime":"2025-12-02T01:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.732773 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.732824 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.732841 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.732866 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.732885 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:30Z","lastTransitionTime":"2025-12-02T01:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.835637 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.835710 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.835759 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.835782 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.835794 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:30Z","lastTransitionTime":"2025-12-02T01:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.938430 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.938504 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.938522 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.938555 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:30 crc kubenswrapper[4884]: I1202 01:37:30.938579 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:30Z","lastTransitionTime":"2025-12-02T01:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.041618 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.041685 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.041707 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.041733 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.041786 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:31Z","lastTransitionTime":"2025-12-02T01:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.144500 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.144542 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.144551 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.144568 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.144581 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:31Z","lastTransitionTime":"2025-12-02T01:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.248080 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.248139 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.248149 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.248168 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.248179 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:31Z","lastTransitionTime":"2025-12-02T01:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.293500 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6l592_791af0d6-a201-430f-b09d-02c24a4b7503/kube-multus/0.log" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.293593 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6l592" event={"ID":"791af0d6-a201-430f-b09d-02c24a4b7503","Type":"ContainerStarted","Data":"b3581bf807ed4bcf2bc452f0ad29348534c8c44bdc2d6a24614c2ea43c6d63a3"} Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.310727 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.334975 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.349346 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.360823 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.360874 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.360892 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.360920 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.360939 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:31Z","lastTransitionTime":"2025-12-02T01:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.391975 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.409941 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.440607 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:08Z\\\",\\\"message\\\":\\\"t:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00013561f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1202 01:37:08.550660 6529 model_client.go:382] Update operations generated as: [{Op:update Table:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:37:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.455183 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aafc3731-36a9-4f87-ba3a-3966b90bd016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27f06f9f8e8f076b46dd2ec9d609217fa18bc94c8f2cc7bf6a6e530820f74e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21ab31a5e5a2b1ae8b6b2d1ee1dcc1aefcc6efe942947fa71b801f90dc346f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a218e1965e88c816e5dc27a6bfe7c6043522794f68c4eca80982082c1c3b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.463546 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.463611 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.463631 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.463659 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.463679 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:31Z","lastTransitionTime":"2025-12-02T01:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.473951 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.493820 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.509978 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.531220 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3581bf807ed4bcf2bc452f0ad29348534c8c44bdc2d6a24614c2ea43c6d63a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:29Z\\\",\\\"message\\\":\\\"2025-12-02T01:36:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61daf367-6f4b-4e21-81a0-cb44a5376679\\\\n2025-12-02T01:36:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61daf367-6f4b-4e21-81a0-cb44a5376679 to /host/opt/cni/bin/\\\\n2025-12-02T01:36:44Z [verbose] multus-daemon started\\\\n2025-12-02T01:36:44Z [verbose] Readiness Indicator file check\\\\n2025-12-02T01:37:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:37:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.549739 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.567619 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.567892 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.567926 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.567959 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.567985 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:31Z","lastTransitionTime":"2025-12-02T01:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.571512 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.587367 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.620697 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.638940 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.654700 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.671334 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.671383 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.671394 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.671414 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.671428 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:31Z","lastTransitionTime":"2025-12-02T01:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.671332 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:31Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.775008 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.775059 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.775070 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.775088 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.775103 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:31Z","lastTransitionTime":"2025-12-02T01:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.878607 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.878688 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.878707 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.878738 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.878795 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:31Z","lastTransitionTime":"2025-12-02T01:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.982301 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.982369 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.982389 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.982416 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:31 crc kubenswrapper[4884]: I1202 01:37:31.982438 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:31Z","lastTransitionTime":"2025-12-02T01:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.084943 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.085005 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.085021 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.085047 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.085066 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:32Z","lastTransitionTime":"2025-12-02T01:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.188475 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.188536 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.188552 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.188577 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.188592 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:32Z","lastTransitionTime":"2025-12-02T01:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.291635 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.291969 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.291989 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.292016 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.292033 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:32Z","lastTransitionTime":"2025-12-02T01:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.394927 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.395006 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.395023 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.395045 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.395058 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:32Z","lastTransitionTime":"2025-12-02T01:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.497910 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.497971 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.497983 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.498002 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.498013 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:32Z","lastTransitionTime":"2025-12-02T01:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.600372 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.600424 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.600436 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.600454 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.600467 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:32Z","lastTransitionTime":"2025-12-02T01:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.613211 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.613305 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.613484 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:32 crc kubenswrapper[4884]: E1202 01:37:32.613565 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.613577 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:32 crc kubenswrapper[4884]: E1202 01:37:32.613322 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:32 crc kubenswrapper[4884]: E1202 01:37:32.613678 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:32 crc kubenswrapper[4884]: E1202 01:37:32.613806 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.702758 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.702798 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.702812 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.702831 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.702844 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:32Z","lastTransitionTime":"2025-12-02T01:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.806118 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.806182 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.806215 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.806232 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.806243 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:32Z","lastTransitionTime":"2025-12-02T01:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.908826 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.908861 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.908870 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.908884 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:32 crc kubenswrapper[4884]: I1202 01:37:32.908895 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:32Z","lastTransitionTime":"2025-12-02T01:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.011958 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.012019 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.012037 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.012062 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.012082 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:33Z","lastTransitionTime":"2025-12-02T01:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.115507 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.115566 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.115582 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.115632 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.115651 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:33Z","lastTransitionTime":"2025-12-02T01:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.218459 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.218516 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.218538 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.218566 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.218589 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:33Z","lastTransitionTime":"2025-12-02T01:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.342410 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.342561 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.342587 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.342610 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.342627 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:33Z","lastTransitionTime":"2025-12-02T01:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.446146 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.446203 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.446219 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.446242 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.446258 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:33Z","lastTransitionTime":"2025-12-02T01:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.549466 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.549530 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.549547 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.549573 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.549591 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:33Z","lastTransitionTime":"2025-12-02T01:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.637468 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.652963 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.653015 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.653100 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.653179 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.653220 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:33Z","lastTransitionTime":"2025-12-02T01:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.655125 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aafc3731-36a9-4f87-ba3a-3966b90bd016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27f06f9f8e8f076b46dd2ec9d609217fa18bc94c8f2cc7bf6a6e530820f74e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21ab31a5e5a2b1ae8b6b2d1ee1dcc1aefcc6efe942947fa71b801f90dc346f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a218e1965e88c816e5dc27a6bfe7c6043522794f68c4eca80982082c1c3b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.686231 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.703814 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.718573 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.735673 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3581bf807ed4bcf2bc452f0ad29348534c8c44bdc2d6a24614c2ea43c6d63a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:29Z\\\",\\\"message\\\":\\\"2025-12-02T01:36:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61daf367-6f4b-4e21-81a0-cb44a5376679\\\\n2025-12-02T01:36:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61daf367-6f4b-4e21-81a0-cb44a5376679 to /host/opt/cni/bin/\\\\n2025-12-02T01:36:44Z [verbose] multus-daemon started\\\\n2025-12-02T01:36:44Z [verbose] Readiness Indicator file check\\\\n2025-12-02T01:37:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:37:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.747818 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.755527 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.755606 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.755624 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.755649 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.755667 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:33Z","lastTransitionTime":"2025-12-02T01:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.766513 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.785514 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.801700 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.812903 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.827367 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.845520 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.858473 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.858510 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.858519 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.858533 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.858543 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:33Z","lastTransitionTime":"2025-12-02T01:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.861902 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.872724 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.892178 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.903965 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.931519 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:08Z\\\",\\\"message\\\":\\\"t:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00013561f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1202 01:37:08.550660 6529 model_client.go:382] Update operations generated as: [{Op:update Table:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:37:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:33Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.960172 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.960212 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.960230 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.960252 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:33 crc kubenswrapper[4884]: I1202 01:37:33.960269 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:33Z","lastTransitionTime":"2025-12-02T01:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.062322 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.062409 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.062443 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.062472 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.062492 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:34Z","lastTransitionTime":"2025-12-02T01:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.164348 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.164402 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.164414 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.164433 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.164445 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:34Z","lastTransitionTime":"2025-12-02T01:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.268299 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.268351 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.268363 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.268383 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.268394 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:34Z","lastTransitionTime":"2025-12-02T01:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.371025 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.371059 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.371067 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.371080 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.371091 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:34Z","lastTransitionTime":"2025-12-02T01:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.473288 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.473348 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.473366 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.473389 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.473411 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:34Z","lastTransitionTime":"2025-12-02T01:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.576315 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.576357 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.576366 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.576382 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.576392 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:34Z","lastTransitionTime":"2025-12-02T01:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.614011 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.614151 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:34 crc kubenswrapper[4884]: E1202 01:37:34.614250 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.614264 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.614321 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:34 crc kubenswrapper[4884]: E1202 01:37:34.614488 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:34 crc kubenswrapper[4884]: E1202 01:37:34.614610 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:34 crc kubenswrapper[4884]: E1202 01:37:34.614782 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.678929 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.678961 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.678971 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.678985 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.678995 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:34Z","lastTransitionTime":"2025-12-02T01:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.781508 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.781572 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.781589 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.781616 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.781642 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:34Z","lastTransitionTime":"2025-12-02T01:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.883634 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.883684 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.883693 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.883707 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.883715 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:34Z","lastTransitionTime":"2025-12-02T01:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.986480 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.986537 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.986548 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.986563 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:34 crc kubenswrapper[4884]: I1202 01:37:34.986573 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:34Z","lastTransitionTime":"2025-12-02T01:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.088835 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.088893 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.088909 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.088935 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.088953 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:35Z","lastTransitionTime":"2025-12-02T01:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.191174 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.191243 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.191265 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.191297 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.191326 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:35Z","lastTransitionTime":"2025-12-02T01:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.294218 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.294252 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.294260 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.294274 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.294283 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:35Z","lastTransitionTime":"2025-12-02T01:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.408858 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.408906 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.408916 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.408932 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.408944 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:35Z","lastTransitionTime":"2025-12-02T01:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.512061 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.512107 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.512119 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.512140 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.512161 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:35Z","lastTransitionTime":"2025-12-02T01:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.615308 4884 scope.go:117] "RemoveContainer" containerID="10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.620381 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.620586 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.620599 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.620613 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.620625 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:35Z","lastTransitionTime":"2025-12-02T01:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.723685 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.723777 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.723788 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.723804 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.723814 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:35Z","lastTransitionTime":"2025-12-02T01:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.826129 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.826168 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.826177 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.826191 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.826201 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:35Z","lastTransitionTime":"2025-12-02T01:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.927788 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.927839 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.927848 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.927863 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:35 crc kubenswrapper[4884]: I1202 01:37:35.927873 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:35Z","lastTransitionTime":"2025-12-02T01:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.030989 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.031076 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.031091 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.031111 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.031126 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:36Z","lastTransitionTime":"2025-12-02T01:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.133638 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.133690 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.133708 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.133729 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.133773 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:36Z","lastTransitionTime":"2025-12-02T01:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.237273 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.237346 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.237369 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.237395 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.237418 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:36Z","lastTransitionTime":"2025-12-02T01:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.311763 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/2.log" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.315379 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerStarted","Data":"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9"} Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.340213 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.340256 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.340275 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.340301 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.340323 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:36Z","lastTransitionTime":"2025-12-02T01:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.445713 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.445765 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.445778 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.445950 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.445962 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:36Z","lastTransitionTime":"2025-12-02T01:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.548307 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.548347 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.548360 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.548375 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.548386 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:36Z","lastTransitionTime":"2025-12-02T01:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.613690 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.613775 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.613790 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.613791 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:36 crc kubenswrapper[4884]: E1202 01:37:36.613887 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:36 crc kubenswrapper[4884]: E1202 01:37:36.614031 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:36 crc kubenswrapper[4884]: E1202 01:37:36.614201 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:36 crc kubenswrapper[4884]: E1202 01:37:36.614311 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.650204 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.650252 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.650264 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.650286 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.650301 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:36Z","lastTransitionTime":"2025-12-02T01:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.752902 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.752957 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.752973 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.752996 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.753012 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:36Z","lastTransitionTime":"2025-12-02T01:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.856006 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.856047 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.856063 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.856083 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.856099 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:36Z","lastTransitionTime":"2025-12-02T01:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.957714 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.957758 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.957767 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.957779 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:36 crc kubenswrapper[4884]: I1202 01:37:36.957788 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:36Z","lastTransitionTime":"2025-12-02T01:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.059839 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.059869 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.059878 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.059892 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.059900 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:37Z","lastTransitionTime":"2025-12-02T01:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.161885 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.161945 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.161967 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.161993 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.162017 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:37Z","lastTransitionTime":"2025-12-02T01:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.264317 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.264379 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.264402 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.264431 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.264452 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:37Z","lastTransitionTime":"2025-12-02T01:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.320612 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.339729 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aafc3731-36a9-4f87-ba3a-3966b90bd016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27f06f9f8e8f076b46dd2ec9d609217fa18bc94c8f2cc7bf6a6e530820f74e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21ab31a5e5a2b1ae8b6b2d1ee1dcc1aefcc6efe942947fa71b801f90dc346f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a218e1965e88c816e5dc27a6bfe7c6043522794f68c4eca80982082c1c3b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.357520 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.370990 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.371037 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.371054 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.371076 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.371094 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:37Z","lastTransitionTime":"2025-12-02T01:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.381297 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.397074 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.414997 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3581bf807ed4bcf2bc452f0ad29348534c8c44bdc2d6a24614c2ea43c6d63a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:29Z\\\",\\\"message\\\":\\\"2025-12-02T01:36:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61daf367-6f4b-4e21-81a0-cb44a5376679\\\\n2025-12-02T01:36:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61daf367-6f4b-4e21-81a0-cb44a5376679 to /host/opt/cni/bin/\\\\n2025-12-02T01:36:44Z [verbose] multus-daemon started\\\\n2025-12-02T01:36:44Z [verbose] Readiness Indicator file check\\\\n2025-12-02T01:37:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:37:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.431647 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.449104 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.463607 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.477242 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.477308 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.477491 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.477519 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.477551 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.477576 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:37Z","lastTransitionTime":"2025-12-02T01:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.490983 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.509693 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.523392 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.539934 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.558128 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.576856 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.580506 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.580549 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.580567 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.580591 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.580611 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:37Z","lastTransitionTime":"2025-12-02T01:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.604500 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.624140 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.650038 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:08Z\\\",\\\"message\\\":\\\"t:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00013561f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1202 01:37:08.550660 6529 model_client.go:382] Update operations generated as: [{Op:update Table:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:37:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:37Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.682501 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.682540 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.682549 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.682563 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.682574 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:37Z","lastTransitionTime":"2025-12-02T01:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.784772 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.784829 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.784846 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.784872 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.784891 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:37Z","lastTransitionTime":"2025-12-02T01:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.887376 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.887436 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.887455 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.887520 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.887540 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:37Z","lastTransitionTime":"2025-12-02T01:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.990338 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.990394 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.990411 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.990436 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:37 crc kubenswrapper[4884]: I1202 01:37:37.990453 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:37Z","lastTransitionTime":"2025-12-02T01:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.010325 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.010365 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.010420 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.010441 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.010456 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:38Z","lastTransitionTime":"2025-12-02T01:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:38 crc kubenswrapper[4884]: E1202 01:37:38.030179 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.034641 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.034769 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.034792 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.034812 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.034828 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:38Z","lastTransitionTime":"2025-12-02T01:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:38 crc kubenswrapper[4884]: E1202 01:37:38.054654 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.058844 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.058904 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.058923 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.058947 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.058966 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:38Z","lastTransitionTime":"2025-12-02T01:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:38 crc kubenswrapper[4884]: E1202 01:37:38.078693 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.084122 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.084171 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.084192 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.084214 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.084230 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:38Z","lastTransitionTime":"2025-12-02T01:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:38 crc kubenswrapper[4884]: E1202 01:37:38.104165 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.108699 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.108776 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.108794 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.108821 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.108839 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:38Z","lastTransitionTime":"2025-12-02T01:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:38 crc kubenswrapper[4884]: E1202 01:37:38.128952 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: E1202 01:37:38.129336 4884 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.131625 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.131678 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.131696 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.131719 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.131736 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:38Z","lastTransitionTime":"2025-12-02T01:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.235205 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.235255 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.235272 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.235298 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.235316 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:38Z","lastTransitionTime":"2025-12-02T01:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.326053 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/3.log" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.327008 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/2.log" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.331097 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8c50c39-2888-444e-bf55-49b091559978" containerID="5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9" exitCode=1 Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.331153 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerDied","Data":"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9"} Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.331201 4884 scope.go:117] "RemoveContainer" containerID="10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.332358 4884 scope.go:117] "RemoveContainer" containerID="5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9" Dec 02 01:37:38 crc kubenswrapper[4884]: E1202 01:37:38.333944 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.337972 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.338014 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.338030 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.338053 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.338070 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:38Z","lastTransitionTime":"2025-12-02T01:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.354610 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.378581 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.397992 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.416712 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.434456 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.441027 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.441100 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.441119 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.441560 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.441619 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:38Z","lastTransitionTime":"2025-12-02T01:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.455703 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.476270 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.494524 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.510948 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.543668 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.550772 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.550960 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.551129 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.551281 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.551443 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:38Z","lastTransitionTime":"2025-12-02T01:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.563655 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.616969 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.617276 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.617060 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.617068 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:38 crc kubenswrapper[4884]: E1202 01:37:38.617839 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:38 crc kubenswrapper[4884]: E1202 01:37:38.617988 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:38 crc kubenswrapper[4884]: E1202 01:37:38.618150 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:38 crc kubenswrapper[4884]: E1202 01:37:38.618368 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.623866 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://10f13c9eeb53fe76be19f1782e77afa1faf47449f1be6af4ff89a75873c74f88\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:08Z\\\",\\\"message\\\":\\\"t:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00013561f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:https,Protocol:TCP,Port:443,TargetPort:{0 8443 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: console-operator,},ClusterIP:10.217.4.88,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.4.88],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1202 01:37:08.550660 6529 model_client.go:382] Update operations generated as: [{Op:update Table:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:37:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:37Z\\\",\\\"message\\\":\\\"rvices.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.37\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 01:37:37.614015 6871 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1202 01:37:37.614024 6871 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1202 01:37:37.613965 6871 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1202 01:37:37.613016 6871 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization,\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:37:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.647857 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.654186 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.654455 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.654584 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.654860 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.655024 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:38Z","lastTransitionTime":"2025-12-02T01:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.667108 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aafc3731-36a9-4f87-ba3a-3966b90bd016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27f06f9f8e8f076b46dd2ec9d609217fa18bc94c8f2cc7bf6a6e530820f74e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21ab31a5e5a2b1ae8b6b2d1ee1dcc1aefcc6efe942947fa71b801f90dc346f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a218e1965e88c816e5dc27a6bfe7c6043522794f68c4eca80982082c1c3b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.688236 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.709791 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.727650 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.748144 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3581bf807ed4bcf2bc452f0ad29348534c8c44bdc2d6a24614c2ea43c6d63a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:29Z\\\",\\\"message\\\":\\\"2025-12-02T01:36:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61daf367-6f4b-4e21-81a0-cb44a5376679\\\\n2025-12-02T01:36:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61daf367-6f4b-4e21-81a0-cb44a5376679 to /host/opt/cni/bin/\\\\n2025-12-02T01:36:44Z [verbose] multus-daemon started\\\\n2025-12-02T01:36:44Z [verbose] Readiness Indicator file check\\\\n2025-12-02T01:37:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:37:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:38Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.757641 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.757714 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.757732 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.757782 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.757804 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:38Z","lastTransitionTime":"2025-12-02T01:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.860959 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.861280 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.861389 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.861495 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.861655 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:38Z","lastTransitionTime":"2025-12-02T01:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.964090 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.964144 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.964160 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.964184 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:38 crc kubenswrapper[4884]: I1202 01:37:38.964202 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:38Z","lastTransitionTime":"2025-12-02T01:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.067429 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.067495 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.067516 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.067544 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.067566 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:39Z","lastTransitionTime":"2025-12-02T01:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.170717 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.170807 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.170830 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.170859 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.170883 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:39Z","lastTransitionTime":"2025-12-02T01:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.274171 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.274248 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.274266 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.274290 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.274307 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:39Z","lastTransitionTime":"2025-12-02T01:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.337878 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/3.log" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.342326 4884 scope.go:117] "RemoveContainer" containerID="5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9" Dec 02 01:37:39 crc kubenswrapper[4884]: E1202 01:37:39.342575 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.363271 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.376666 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.376726 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.376770 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.376798 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.376818 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:39Z","lastTransitionTime":"2025-12-02T01:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.383828 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.401309 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.433903 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.453995 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.480138 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.480192 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.480210 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.480235 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.480253 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:39Z","lastTransitionTime":"2025-12-02T01:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.484836 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:37Z\\\",\\\"message\\\":\\\"rvices.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.37\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 01:37:37.614015 6871 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1202 01:37:37.614024 6871 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1202 01:37:37.613965 6871 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1202 01:37:37.613016 6871 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization,\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:37:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.502728 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aafc3731-36a9-4f87-ba3a-3966b90bd016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27f06f9f8e8f076b46dd2ec9d609217fa18bc94c8f2cc7bf6a6e530820f74e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21ab31a5e5a2b1ae8b6b2d1ee1dcc1aefcc6efe942947fa71b801f90dc346f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a218e1965e88c816e5dc27a6bfe7c6043522794f68c4eca80982082c1c3b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.522604 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.541940 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.559025 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.581873 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3581bf807ed4bcf2bc452f0ad29348534c8c44bdc2d6a24614c2ea43c6d63a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:29Z\\\",\\\"message\\\":\\\"2025-12-02T01:36:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61daf367-6f4b-4e21-81a0-cb44a5376679\\\\n2025-12-02T01:36:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61daf367-6f4b-4e21-81a0-cb44a5376679 to /host/opt/cni/bin/\\\\n2025-12-02T01:36:44Z [verbose] multus-daemon started\\\\n2025-12-02T01:36:44Z [verbose] Readiness Indicator file check\\\\n2025-12-02T01:37:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:37:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.583415 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.583478 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.583500 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.583528 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.583550 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:39Z","lastTransitionTime":"2025-12-02T01:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.604832 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.626827 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.653996 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.675104 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.686430 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.686494 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.686518 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.686546 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.686570 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:39Z","lastTransitionTime":"2025-12-02T01:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.695230 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.712907 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.729460 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:39Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.789351 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.789406 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.789425 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.789447 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.789465 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:39Z","lastTransitionTime":"2025-12-02T01:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.892842 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.892938 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.892955 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.892977 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.892996 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:39Z","lastTransitionTime":"2025-12-02T01:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.995341 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.995400 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.995416 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.995439 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:39 crc kubenswrapper[4884]: I1202 01:37:39.995457 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:39Z","lastTransitionTime":"2025-12-02T01:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.098352 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.098422 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.098442 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.098466 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.098483 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:40Z","lastTransitionTime":"2025-12-02T01:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.202098 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.202157 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.202172 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.202192 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.202207 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:40Z","lastTransitionTime":"2025-12-02T01:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.304831 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.304889 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.304906 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.305124 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.305144 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:40Z","lastTransitionTime":"2025-12-02T01:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.408340 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.408402 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.408421 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.408445 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.408463 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:40Z","lastTransitionTime":"2025-12-02T01:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.511722 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.511801 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.511819 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.511847 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.511864 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:40Z","lastTransitionTime":"2025-12-02T01:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.613108 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.613114 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:40 crc kubenswrapper[4884]: E1202 01:37:40.613327 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.613418 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.613418 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:40 crc kubenswrapper[4884]: E1202 01:37:40.613580 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:40 crc kubenswrapper[4884]: E1202 01:37:40.613664 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:40 crc kubenswrapper[4884]: E1202 01:37:40.613792 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.615033 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.615092 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.615117 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.615144 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.615167 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:40Z","lastTransitionTime":"2025-12-02T01:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.718584 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.718648 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.718671 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.718702 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.718724 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:40Z","lastTransitionTime":"2025-12-02T01:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.823007 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.823080 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.823103 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.823132 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.823157 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:40Z","lastTransitionTime":"2025-12-02T01:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.926662 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.926718 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.926985 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.927043 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:40 crc kubenswrapper[4884]: I1202 01:37:40.927068 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:40Z","lastTransitionTime":"2025-12-02T01:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.030686 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.030844 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.030866 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.030903 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.030926 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:41Z","lastTransitionTime":"2025-12-02T01:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.134305 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.134371 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.134392 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.134420 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.134439 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:41Z","lastTransitionTime":"2025-12-02T01:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.237604 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.237663 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.237680 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.237710 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.237729 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:41Z","lastTransitionTime":"2025-12-02T01:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.341714 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.341805 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.341826 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.341854 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.341872 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:41Z","lastTransitionTime":"2025-12-02T01:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.444589 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.444649 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.444671 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.444701 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.444723 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:41Z","lastTransitionTime":"2025-12-02T01:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.548218 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.548257 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.548266 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.548279 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.548288 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:41Z","lastTransitionTime":"2025-12-02T01:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.651547 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.651599 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.651610 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.651628 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.651642 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:41Z","lastTransitionTime":"2025-12-02T01:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.755359 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.755413 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.755430 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.755452 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.755467 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:41Z","lastTransitionTime":"2025-12-02T01:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.857867 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.857932 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.857948 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.857971 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.857987 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:41Z","lastTransitionTime":"2025-12-02T01:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.961797 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.961999 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.962021 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.962047 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:41 crc kubenswrapper[4884]: I1202 01:37:41.962065 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:41Z","lastTransitionTime":"2025-12-02T01:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.064788 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.064839 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.064857 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.064881 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.064899 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:42Z","lastTransitionTime":"2025-12-02T01:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.167378 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.167447 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.167465 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.167490 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.167508 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:42Z","lastTransitionTime":"2025-12-02T01:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.270831 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.270889 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.270908 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.270933 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.270951 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:42Z","lastTransitionTime":"2025-12-02T01:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.373670 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.373740 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.373813 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.373843 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.373867 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:42Z","lastTransitionTime":"2025-12-02T01:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.477284 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.477361 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.477388 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.477420 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.477441 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:42Z","lastTransitionTime":"2025-12-02T01:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.580305 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.580351 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.580368 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.580390 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.580407 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:42Z","lastTransitionTime":"2025-12-02T01:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.613736 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:42 crc kubenswrapper[4884]: E1202 01:37:42.613999 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.614393 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:42 crc kubenswrapper[4884]: E1202 01:37:42.614551 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.614875 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:42 crc kubenswrapper[4884]: E1202 01:37:42.615021 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.615300 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:42 crc kubenswrapper[4884]: E1202 01:37:42.615459 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.683612 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.683667 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.683687 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.683712 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.683731 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:42Z","lastTransitionTime":"2025-12-02T01:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.785926 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.785992 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.786011 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.786038 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.786056 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:42Z","lastTransitionTime":"2025-12-02T01:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.889261 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.889324 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.889342 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.889365 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.889385 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:42Z","lastTransitionTime":"2025-12-02T01:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.992072 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.992127 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.992146 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.992169 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:42 crc kubenswrapper[4884]: I1202 01:37:42.992186 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:42Z","lastTransitionTime":"2025-12-02T01:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.094726 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.094788 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.094803 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.094821 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.094835 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:43Z","lastTransitionTime":"2025-12-02T01:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.198273 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.198431 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.198488 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.198519 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.198545 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:43Z","lastTransitionTime":"2025-12-02T01:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.301054 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.301144 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.301439 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.301832 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.302064 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:43Z","lastTransitionTime":"2025-12-02T01:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.404360 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.404417 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.404435 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.404463 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.404482 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:43Z","lastTransitionTime":"2025-12-02T01:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.508140 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.508290 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.508306 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.508329 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.508346 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:43Z","lastTransitionTime":"2025-12-02T01:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.611952 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.612034 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.612058 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.612091 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.612113 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:43Z","lastTransitionTime":"2025-12-02T01:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.676462 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-p545z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"531c7e95-83b1-4ad1-ad0e-490d9520e8d6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f6ed2e8617042efdac0dc1cf3025b14f0e73e136272105ea0c2dbdaaeefdc1d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c824x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-p545z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.697083 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://76228d5c31dc3cbf30b7c27f9b34bdae60bd1199a0dd6d0639244e40f8fb994a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22d3c8ae8344c0ecc7c8788017707aabf6a797ef96893ab5e487218563be4eec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.710551 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5313d76d92b3c642f9ca1b11c4c9f1180fad6b5d3b3f93c6a6ebc22e47761e02\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.714845 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.714889 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.714904 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.714925 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.714940 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:43Z","lastTransitionTime":"2025-12-02T01:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.733847 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b1a7366c-fc1b-405a-b3b7-b48b414b1d48\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6c1f443a39bea78ce45a5facdb6b7757469923b0b12a5b2491b22a169bad5c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c776510cd206dc4f69eca08de7b31eb3d6fc6a83d9a7371d29a6639576054f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c13ed6a27864223f7f4d508c4d31c4cb2134ded31b4d41cc61df9fc8c8f9c77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1975e42e68e6fa10ee3b04c113cb08399e707d2e47acad86c3403cad4d4902d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://344079845fa745db69e2f23c653fac81cf99f5149314c0c55ffc55222f4bc6e4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2f8823c128cedfa576531163342f6d9c80e97c47d630657bb2ec1dd423322ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a01aa87f720abc756af72bffa1689057f4ac2b4252a10538b52e065163688a36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b328ebfcc77611d7e9a75522dfdabb89e7a8ca432ca16c8573bc9b5b4f8d370f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.748906 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.772260 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d8c50c39-2888-444e-bf55-49b091559978\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:37Z\\\",\\\"message\\\":\\\"rvices.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.37\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1202 01:37:37.614015 6871 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI1202 01:37:37.614024 6871 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI1202 01:37:37.613965 6871 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]} options:{GoMap:map[iface-id-ver:3b6479f0-333b-4a96-9adf-2099afdc2447 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:04 10.217.0.4]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {61897e97-c771-4738-8709-09636387cb00}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1202 01:37:37.613016 6871 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization,\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:37:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bx2hb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-c74x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.788923 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9385a5c9-a156-460a-8ca1-9b342ab835c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e5caa7449c520868b70a7016ac97c9c81e55eb715a14ea3af53240d696bf9ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hwhc8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-jc7vx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.807835 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6l592" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"791af0d6-a201-430f-b09d-02c24a4b7503\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3581bf807ed4bcf2bc452f0ad29348534c8c44bdc2d6a24614c2ea43c6d63a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-02T01:37:29Z\\\",\\\"message\\\":\\\"2025-12-02T01:36:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_61daf367-6f4b-4e21-81a0-cb44a5376679\\\\n2025-12-02T01:36:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_61daf367-6f4b-4e21-81a0-cb44a5376679 to /host/opt/cni/bin/\\\\n2025-12-02T01:36:44Z [verbose] multus-daemon started\\\\n2025-12-02T01:36:44Z [verbose] Readiness Indicator file check\\\\n2025-12-02T01:37:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:37:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7xcl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6l592\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.818034 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.818084 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.818098 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.818117 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.818131 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:43Z","lastTransitionTime":"2025-12-02T01:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.830386 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32059fa0-1cad-40d4-92ac-2ee9e89ed661\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://412c9a82ec7725affaad5c2fb42dda7fe3ac82e26a965840e9fe1d2e6886b7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7c66e54efa2996ea0c7d45694c8afffb77dde260dd803268db237feeafbe577e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb064b2c4b9c6c3fca4cf021465a6f72b37355708731f66acef188789ea327a6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f037a9e3f93621039c8ee0a40a4631b19d3d191d443c70e8a80bfad28613a93f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0339b5463da02927b4dd13195c980c4e89602535423d008d0e16848235c0fb69\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2199a720cfc1f5d939cbf5c3314ed2b5d9e5c262e8851c0f0c0db13de2d22046\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://84f77c6197cc4d93ad8a37b5544385a769766a12306e504587fd5a2747f04639\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ghczf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mlkwv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.846964 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aafc3731-36a9-4f87-ba3a-3966b90bd016\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27f06f9f8e8f076b46dd2ec9d609217fa18bc94c8f2cc7bf6a6e530820f74e92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21ab31a5e5a2b1ae8b6b2d1ee1dcc1aefcc6efe942947fa71b801f90dc346f1f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67a218e1965e88c816e5dc27a6bfe7c6043522794f68c4eca80982082c1c3b62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://87c9f6df6b31230de756e0b8fd9126644dbce0438be05691fc8c3db9e3915353\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.868092 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:43Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22a9f38516e83aa98f60c1daf548ef1895c980ffb630e02e598dfcf5880c7895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.883682 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.898875 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.915810 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.920198 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.920230 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.920242 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.920260 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.920274 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:43Z","lastTransitionTime":"2025-12-02T01:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.927092 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.944983 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdba8a22-47be-4221-b923-a41afee4989d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1202 01:36:36.620042 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1202 01:36:36.622086 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-819670676/tls.crt::/tmp/serving-cert-819670676/tls.key\\\\\\\"\\\\nI1202 01:36:42.124722 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1202 01:36:42.128459 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1202 01:36:42.128494 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1202 01:36:42.128520 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1202 01:36:42.128527 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1202 01:36:42.152879 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1202 01:36:42.152906 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152911 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1202 01:36:42.152918 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1202 01:36:42.152921 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1202 01:36:42.152926 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1202 01:36:42.152928 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1202 01:36:42.152974 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1202 01:36:42.161804 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-02T01:36:24Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.963291 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8fec2cbd-6893-4a45-a66b-6781a55df9bb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69fbef2c40d3c128c4a5d7afdcfb062ed548d35942f858e5eced9527f9c4d69a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a001f8b04b758b4bc4a71338497f79433d3cf8f3747fd81e187add1be3b938e0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec96f640431b11ef182607d90c833e4b1e679de88569ead3d9f20e09779df8f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:23Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:43 crc kubenswrapper[4884]: I1202 01:37:43.978967 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:43Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.023010 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.023064 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.023081 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.023107 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.023125 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:44Z","lastTransitionTime":"2025-12-02T01:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.126217 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.126587 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.126606 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.126631 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.126649 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:44Z","lastTransitionTime":"2025-12-02T01:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.230172 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.230224 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.230242 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.230267 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.230285 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:44Z","lastTransitionTime":"2025-12-02T01:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.332673 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.332731 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.332784 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.332807 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.332826 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:44Z","lastTransitionTime":"2025-12-02T01:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.435700 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.435774 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.435791 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.435811 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.435827 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:44Z","lastTransitionTime":"2025-12-02T01:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.538201 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.538262 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.538280 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.538304 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.538321 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:44Z","lastTransitionTime":"2025-12-02T01:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.613687 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.613794 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.613839 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.613691 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:44 crc kubenswrapper[4884]: E1202 01:37:44.613950 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:44 crc kubenswrapper[4884]: E1202 01:37:44.614180 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:44 crc kubenswrapper[4884]: E1202 01:37:44.614233 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:44 crc kubenswrapper[4884]: E1202 01:37:44.614457 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.641372 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.641428 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.641522 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.641551 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.641604 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:44Z","lastTransitionTime":"2025-12-02T01:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.744225 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.744283 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.744304 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.744332 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.744350 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:44Z","lastTransitionTime":"2025-12-02T01:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.847857 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.847912 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.847930 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.847954 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.847972 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:44Z","lastTransitionTime":"2025-12-02T01:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.950872 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.950983 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.951008 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.951034 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:44 crc kubenswrapper[4884]: I1202 01:37:44.951056 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:44Z","lastTransitionTime":"2025-12-02T01:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.054393 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.054454 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.054470 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.054524 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.054541 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:45Z","lastTransitionTime":"2025-12-02T01:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.157871 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.157926 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.157949 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.157973 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.157991 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:45Z","lastTransitionTime":"2025-12-02T01:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.260618 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.260681 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.260690 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.260733 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.260761 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:45Z","lastTransitionTime":"2025-12-02T01:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.363589 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.363696 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.363716 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.363735 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.363796 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:45Z","lastTransitionTime":"2025-12-02T01:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.467040 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.467102 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.467120 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.467147 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.467167 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:45Z","lastTransitionTime":"2025-12-02T01:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.569823 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.569916 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.569933 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.569957 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.569974 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:45Z","lastTransitionTime":"2025-12-02T01:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.673779 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.673840 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.673899 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.673931 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.673991 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:45Z","lastTransitionTime":"2025-12-02T01:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.776951 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.777131 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.777160 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.777192 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.777215 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:45Z","lastTransitionTime":"2025-12-02T01:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.880333 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.880396 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.880413 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.880437 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.880455 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:45Z","lastTransitionTime":"2025-12-02T01:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.984650 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.984721 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.984789 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.984823 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:45 crc kubenswrapper[4884]: I1202 01:37:45.984845 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:45Z","lastTransitionTime":"2025-12-02T01:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.087961 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.088014 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.088030 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.088053 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.088070 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:46Z","lastTransitionTime":"2025-12-02T01:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.190938 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.190991 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.191007 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.191032 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.191049 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:46Z","lastTransitionTime":"2025-12-02T01:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.294185 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.294244 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.294261 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.294285 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.294302 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:46Z","lastTransitionTime":"2025-12-02T01:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.397056 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.397107 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.397118 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.397135 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.397147 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:46Z","lastTransitionTime":"2025-12-02T01:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.500530 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.500586 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.500604 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.500631 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.500648 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:46Z","lastTransitionTime":"2025-12-02T01:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.508219 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.508557 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:50.508527738 +0000 UTC m=+147.184364662 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.603985 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.604152 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.604176 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.604200 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.604220 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:46Z","lastTransitionTime":"2025-12-02T01:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.609129 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.609184 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.609206 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.609225 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.609300 4884 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.609349 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:38:50.609336969 +0000 UTC m=+147.285173853 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.609481 4884 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.609527 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.609562 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.609574 4884 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.609591 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-02 01:38:50.609557315 +0000 UTC m=+147.285394239 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.609628 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-02 01:38:50.609609376 +0000 UTC m=+147.285446260 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.609489 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.609687 4884 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.609713 4884 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.609841 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-02 01:38:50.609816131 +0000 UTC m=+147.285653045 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.613118 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.613296 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.613326 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.613327 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.613492 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.613441 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.613718 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:46 crc kubenswrapper[4884]: E1202 01:37:46.613731 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.707422 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.707474 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.707495 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.707519 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.707537 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:46Z","lastTransitionTime":"2025-12-02T01:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.810924 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.810988 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.811006 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.811033 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.811054 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:46Z","lastTransitionTime":"2025-12-02T01:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.914402 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.914469 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.914491 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.914520 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:46 crc kubenswrapper[4884]: I1202 01:37:46.914540 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:46Z","lastTransitionTime":"2025-12-02T01:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.017579 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.017662 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.017689 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.017715 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.017732 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:47Z","lastTransitionTime":"2025-12-02T01:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.119724 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.119848 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.119870 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.119898 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.119921 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:47Z","lastTransitionTime":"2025-12-02T01:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.222377 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.222454 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.222478 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.222509 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.222531 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:47Z","lastTransitionTime":"2025-12-02T01:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.325567 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.325657 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.325682 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.325711 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.325736 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:47Z","lastTransitionTime":"2025-12-02T01:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.428391 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.428455 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.428472 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.428531 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.428551 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:47Z","lastTransitionTime":"2025-12-02T01:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.531523 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.531602 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.531625 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.531653 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.531673 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:47Z","lastTransitionTime":"2025-12-02T01:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.634913 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.634984 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.635006 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.635037 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.635066 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:47Z","lastTransitionTime":"2025-12-02T01:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.737812 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.737904 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.737930 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.737962 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.737989 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:47Z","lastTransitionTime":"2025-12-02T01:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.841413 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.841469 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.841484 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.841508 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.841525 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:47Z","lastTransitionTime":"2025-12-02T01:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.944822 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.944877 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.944889 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.944906 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:47 crc kubenswrapper[4884]: I1202 01:37:47.944918 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:47Z","lastTransitionTime":"2025-12-02T01:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.048319 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.048396 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.048436 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.048466 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.048486 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:48Z","lastTransitionTime":"2025-12-02T01:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.150997 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.151049 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.151066 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.151088 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.151105 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:48Z","lastTransitionTime":"2025-12-02T01:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.253456 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.253501 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.253513 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.253528 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.253538 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:48Z","lastTransitionTime":"2025-12-02T01:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.351571 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.351626 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.351643 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.351666 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.351684 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:48Z","lastTransitionTime":"2025-12-02T01:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:48 crc kubenswrapper[4884]: E1202 01:37:48.373033 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.378915 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.378979 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.378997 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.379023 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.379048 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:48Z","lastTransitionTime":"2025-12-02T01:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:48 crc kubenswrapper[4884]: E1202 01:37:48.401231 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.406333 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.406398 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.406421 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.406449 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.406476 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:48Z","lastTransitionTime":"2025-12-02T01:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:48 crc kubenswrapper[4884]: E1202 01:37:48.430079 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.434303 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.434359 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.434378 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.434401 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.434418 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:48Z","lastTransitionTime":"2025-12-02T01:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:48 crc kubenswrapper[4884]: E1202 01:37:48.456534 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.460441 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.460480 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.460534 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.460551 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.460562 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:48Z","lastTransitionTime":"2025-12-02T01:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:48 crc kubenswrapper[4884]: E1202 01:37:48.481940 4884 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-02T01:37:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"54af654b-c7be-43b5-9dfc-8a2692aad92d\\\",\\\"systemUUID\\\":\\\"09169ac4-b9cd-4f4b-a571-21877303afe5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:48Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:48 crc kubenswrapper[4884]: E1202 01:37:48.482204 4884 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.484176 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.484235 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.484254 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.484279 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.484298 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:48Z","lastTransitionTime":"2025-12-02T01:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.587087 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.587151 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.587170 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.587200 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.587230 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:48Z","lastTransitionTime":"2025-12-02T01:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.613945 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.613989 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:48 crc kubenswrapper[4884]: E1202 01:37:48.614127 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.614193 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.614202 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:48 crc kubenswrapper[4884]: E1202 01:37:48.614307 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:48 crc kubenswrapper[4884]: E1202 01:37:48.614438 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:48 crc kubenswrapper[4884]: E1202 01:37:48.614555 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.690205 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.690254 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.690270 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.690292 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.690309 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:48Z","lastTransitionTime":"2025-12-02T01:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.793791 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.793855 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.793873 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.793898 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.793917 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:48Z","lastTransitionTime":"2025-12-02T01:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.896996 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.897070 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.897162 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.897264 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:48 crc kubenswrapper[4884]: I1202 01:37:48.897292 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:48Z","lastTransitionTime":"2025-12-02T01:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.000732 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.000823 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.000840 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.000862 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.000879 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:49Z","lastTransitionTime":"2025-12-02T01:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.103526 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.103590 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.103612 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.103641 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.103661 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:49Z","lastTransitionTime":"2025-12-02T01:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.207161 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.207216 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.207234 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.207260 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.207278 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:49Z","lastTransitionTime":"2025-12-02T01:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.310834 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.310904 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.310920 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.311337 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.311569 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:49Z","lastTransitionTime":"2025-12-02T01:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.415858 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.415924 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.415942 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.415965 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.415983 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:49Z","lastTransitionTime":"2025-12-02T01:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.519000 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.519062 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.519080 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.519104 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.519122 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:49Z","lastTransitionTime":"2025-12-02T01:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.622021 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.622089 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.622110 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.622136 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.622159 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:49Z","lastTransitionTime":"2025-12-02T01:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.725398 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.725527 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.725547 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.725567 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.725602 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:49Z","lastTransitionTime":"2025-12-02T01:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.828991 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.829077 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.829101 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.829131 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.829155 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:49Z","lastTransitionTime":"2025-12-02T01:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.932395 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.932480 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.932501 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.932525 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:49 crc kubenswrapper[4884]: I1202 01:37:49.932542 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:49Z","lastTransitionTime":"2025-12-02T01:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.035878 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.036013 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.036040 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.036072 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.036096 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:50Z","lastTransitionTime":"2025-12-02T01:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.139552 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.139625 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.139698 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.139727 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.139770 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:50Z","lastTransitionTime":"2025-12-02T01:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.243328 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.243391 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.243410 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.243439 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.243456 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:50Z","lastTransitionTime":"2025-12-02T01:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.346363 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.346431 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.346449 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.346474 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.346493 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:50Z","lastTransitionTime":"2025-12-02T01:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.450047 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.450099 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.450119 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.450144 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.450162 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:50Z","lastTransitionTime":"2025-12-02T01:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.554490 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.554591 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.554664 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.554698 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.554856 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:50Z","lastTransitionTime":"2025-12-02T01:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.613849 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.613948 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.613961 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.614039 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:50 crc kubenswrapper[4884]: E1202 01:37:50.614223 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:50 crc kubenswrapper[4884]: E1202 01:37:50.614423 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:50 crc kubenswrapper[4884]: E1202 01:37:50.614829 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:50 crc kubenswrapper[4884]: E1202 01:37:50.614909 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.657782 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.657840 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.657858 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.657884 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.657905 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:50Z","lastTransitionTime":"2025-12-02T01:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.761112 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.761173 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.761190 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.761214 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.761233 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:50Z","lastTransitionTime":"2025-12-02T01:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.864109 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.864206 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.864225 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.864251 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.864269 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:50Z","lastTransitionTime":"2025-12-02T01:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.967553 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.967620 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.967639 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.967663 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:50 crc kubenswrapper[4884]: I1202 01:37:50.967681 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:50Z","lastTransitionTime":"2025-12-02T01:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.070120 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.070178 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.070195 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.070220 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.070268 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:51Z","lastTransitionTime":"2025-12-02T01:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.173433 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.173505 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.173525 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.173547 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.173565 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:51Z","lastTransitionTime":"2025-12-02T01:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.276201 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.276262 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.276279 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.276302 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.276319 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:51Z","lastTransitionTime":"2025-12-02T01:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.379271 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.379307 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.379318 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.379332 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.379343 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:51Z","lastTransitionTime":"2025-12-02T01:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.481462 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.481495 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.481504 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.481517 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.481528 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:51Z","lastTransitionTime":"2025-12-02T01:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.584967 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.585042 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.585060 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.585086 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.585104 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:51Z","lastTransitionTime":"2025-12-02T01:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.688336 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.688446 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.688465 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.688521 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.688539 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:51Z","lastTransitionTime":"2025-12-02T01:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.791609 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.791675 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.791694 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.791721 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.791739 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:51Z","lastTransitionTime":"2025-12-02T01:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.895065 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.895181 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.895207 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.895236 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.895257 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:51Z","lastTransitionTime":"2025-12-02T01:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.998602 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.998661 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.998712 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.998736 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:51 crc kubenswrapper[4884]: I1202 01:37:51.998781 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:51Z","lastTransitionTime":"2025-12-02T01:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.102010 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.102257 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.102421 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.102578 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.102708 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:52Z","lastTransitionTime":"2025-12-02T01:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.205916 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.206084 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.206116 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.206150 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.206181 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:52Z","lastTransitionTime":"2025-12-02T01:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.309553 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.309605 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.309628 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.309655 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.309676 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:52Z","lastTransitionTime":"2025-12-02T01:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.412165 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.412258 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.412277 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.412299 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.412357 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:52Z","lastTransitionTime":"2025-12-02T01:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.515415 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.515478 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.515528 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.515561 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.515584 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:52Z","lastTransitionTime":"2025-12-02T01:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.614462 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.614524 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.614574 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.614471 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:52 crc kubenswrapper[4884]: E1202 01:37:52.614669 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:52 crc kubenswrapper[4884]: E1202 01:37:52.615204 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:52 crc kubenswrapper[4884]: E1202 01:37:52.615370 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.615655 4884 scope.go:117] "RemoveContainer" containerID="5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9" Dec 02 01:37:52 crc kubenswrapper[4884]: E1202 01:37:52.615809 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:52 crc kubenswrapper[4884]: E1202 01:37:52.615994 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.619052 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.619117 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.619145 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.620122 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.620199 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:52Z","lastTransitionTime":"2025-12-02T01:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.723423 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.723466 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.723480 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.723498 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.723510 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:52Z","lastTransitionTime":"2025-12-02T01:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.825981 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.826060 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.826084 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.826118 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.826144 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:52Z","lastTransitionTime":"2025-12-02T01:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.929490 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.929547 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.929565 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.929589 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:52 crc kubenswrapper[4884]: I1202 01:37:52.929609 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:52Z","lastTransitionTime":"2025-12-02T01:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.033004 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.033076 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.033093 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.033120 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.033139 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:53Z","lastTransitionTime":"2025-12-02T01:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.136395 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.136451 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.136468 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.136502 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.136537 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:53Z","lastTransitionTime":"2025-12-02T01:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.239124 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.239188 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.239200 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.239216 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.239229 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:53Z","lastTransitionTime":"2025-12-02T01:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.342773 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.342829 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.342840 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.342858 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.342871 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:53Z","lastTransitionTime":"2025-12-02T01:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.445620 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.445695 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.445709 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.445727 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.445754 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:53Z","lastTransitionTime":"2025-12-02T01:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.548033 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.548069 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.548077 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.548091 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.548102 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:53Z","lastTransitionTime":"2025-12-02T01:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.632337 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.633604 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:42Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.651019 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.651073 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.651085 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.651108 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.651120 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:53Z","lastTransitionTime":"2025-12-02T01:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.651788 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ec4bbba9-9a8c-489e-8a21-198f1cb12a57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59265f28fc476ea1214253ac77c4a86acd17f936562fd9d17dbc57250ec8df38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e34bfa2653cbe1fe0d6f0d4deeda1bc4c6145455f5d7178a3df1f93cf1f7768\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pln92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-fdrt4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.665641 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-whh6t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a6a67e3-3b98-45df-9573-6efee410a90c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8wtwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-whh6t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.680626 4884 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-ww4r4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6b8c081-9419-49d2-a77f-c4fc1f136379\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-02T01:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://461c43b881466ff2d818e90c20f7e9e7660358a7af60f45be4f344fca5967ad5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-02T01:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rnr86\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-02T01:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-ww4r4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-02T01:37:53Z is after 2025-08-24T17:21:41Z" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.768668 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.768704 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.768714 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.768732 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.768761 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:53Z","lastTransitionTime":"2025-12-02T01:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.785258 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=71.785238101 podStartE2EDuration="1m11.785238101s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:37:53.766165697 +0000 UTC m=+90.442002591" watchObservedRunningTime="2025-12-02 01:37:53.785238101 +0000 UTC m=+90.461075005" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.785663 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=71.785654431 podStartE2EDuration="1m11.785654431s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:37:53.783912728 +0000 UTC m=+90.459749672" watchObservedRunningTime="2025-12-02 01:37:53.785654431 +0000 UTC m=+90.461491335" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.820615 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-p545z" podStartSLOduration=71.82058225 podStartE2EDuration="1m11.82058225s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:37:53.820324754 +0000 UTC m=+90.496161668" watchObservedRunningTime="2025-12-02 01:37:53.82058225 +0000 UTC m=+90.496419174" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.872173 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.872216 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.872227 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.872242 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.872253 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:53Z","lastTransitionTime":"2025-12-02T01:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.941996 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=70.941976492 podStartE2EDuration="1m10.941976492s" podCreationTimestamp="2025-12-02 01:36:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:37:53.926132467 +0000 UTC m=+90.601969351" watchObservedRunningTime="2025-12-02 01:37:53.941976492 +0000 UTC m=+90.617813376" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.974327 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.974599 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.974696 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.974813 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.974904 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:53Z","lastTransitionTime":"2025-12-02T01:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.993965 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podStartSLOduration=71.993941886 podStartE2EDuration="1m11.993941886s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:37:53.98095178 +0000 UTC m=+90.656788674" watchObservedRunningTime="2025-12-02 01:37:53.993941886 +0000 UTC m=+90.669778770" Dec 02 01:37:53 crc kubenswrapper[4884]: I1202 01:37:53.994322 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-6l592" podStartSLOduration=71.994318195 podStartE2EDuration="1m11.994318195s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:37:53.993826803 +0000 UTC m=+90.669663697" watchObservedRunningTime="2025-12-02 01:37:53.994318195 +0000 UTC m=+90.670155079" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.012351 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-mlkwv" podStartSLOduration=72.012333383 podStartE2EDuration="1m12.012333383s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:37:54.011703058 +0000 UTC m=+90.687539952" watchObservedRunningTime="2025-12-02 01:37:54.012333383 +0000 UTC m=+90.688170277" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.030437 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=44.030423143 podStartE2EDuration="44.030423143s" podCreationTimestamp="2025-12-02 01:37:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:37:54.030387712 +0000 UTC m=+90.706224606" watchObservedRunningTime="2025-12-02 01:37:54.030423143 +0000 UTC m=+90.706260027" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.078204 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.078273 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.078291 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.078316 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.078333 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:54Z","lastTransitionTime":"2025-12-02T01:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.181575 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.181639 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.181658 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.181684 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.181705 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:54Z","lastTransitionTime":"2025-12-02T01:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.285660 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.285724 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.285765 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.285793 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.285812 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:54Z","lastTransitionTime":"2025-12-02T01:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.388517 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.388577 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.388594 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.388618 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.388635 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:54Z","lastTransitionTime":"2025-12-02T01:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.491932 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.492012 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.492031 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.492058 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.492079 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:54Z","lastTransitionTime":"2025-12-02T01:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.595732 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.595826 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.595844 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.595870 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.595889 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:54Z","lastTransitionTime":"2025-12-02T01:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.613197 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.613273 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.613313 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:54 crc kubenswrapper[4884]: E1202 01:37:54.613391 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.613478 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:54 crc kubenswrapper[4884]: E1202 01:37:54.613521 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:54 crc kubenswrapper[4884]: E1202 01:37:54.613791 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:54 crc kubenswrapper[4884]: E1202 01:37:54.613998 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.699064 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.699127 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.699196 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.699226 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.699291 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:54Z","lastTransitionTime":"2025-12-02T01:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.802271 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.802330 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.802347 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.802369 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.802389 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:54Z","lastTransitionTime":"2025-12-02T01:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.905640 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.905711 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.905729 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.905789 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:54 crc kubenswrapper[4884]: I1202 01:37:54.905810 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:54Z","lastTransitionTime":"2025-12-02T01:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.011529 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.011620 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.011635 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.011666 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.011681 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:55Z","lastTransitionTime":"2025-12-02T01:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.115269 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.115335 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.115348 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.115372 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.115389 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:55Z","lastTransitionTime":"2025-12-02T01:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.218310 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.218370 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.218387 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.218408 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.218428 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:55Z","lastTransitionTime":"2025-12-02T01:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.321734 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.321805 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.321817 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.321839 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.321852 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:55Z","lastTransitionTime":"2025-12-02T01:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.425110 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.425172 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.425189 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.425213 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.425234 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:55Z","lastTransitionTime":"2025-12-02T01:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.528412 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.528500 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.528530 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.528556 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.528574 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:55Z","lastTransitionTime":"2025-12-02T01:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.630820 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.630895 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.630914 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.630946 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.630970 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:55Z","lastTransitionTime":"2025-12-02T01:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.739300 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.739385 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.739459 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.739485 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.739503 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:55Z","lastTransitionTime":"2025-12-02T01:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.842660 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.842786 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.842815 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.842845 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.842862 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:55Z","lastTransitionTime":"2025-12-02T01:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.946086 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.946148 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.946166 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.946194 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:55 crc kubenswrapper[4884]: I1202 01:37:55.946215 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:55Z","lastTransitionTime":"2025-12-02T01:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.049217 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.049288 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.049307 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.049332 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.049351 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:56Z","lastTransitionTime":"2025-12-02T01:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.152418 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.152475 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.152492 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.152519 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.152536 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:56Z","lastTransitionTime":"2025-12-02T01:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.256078 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.256135 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.256153 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.256177 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.256194 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:56Z","lastTransitionTime":"2025-12-02T01:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.359621 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.359681 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.359698 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.359724 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.359740 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:56Z","lastTransitionTime":"2025-12-02T01:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.462858 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.463014 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.463036 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.463063 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.463127 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:56Z","lastTransitionTime":"2025-12-02T01:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.565712 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.565843 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.565882 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.565916 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.565939 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:56Z","lastTransitionTime":"2025-12-02T01:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.613447 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.613518 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.613561 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.613580 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:56 crc kubenswrapper[4884]: E1202 01:37:56.613801 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:56 crc kubenswrapper[4884]: E1202 01:37:56.613927 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:56 crc kubenswrapper[4884]: E1202 01:37:56.614066 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:56 crc kubenswrapper[4884]: E1202 01:37:56.614195 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.668645 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.668686 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.668697 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.668717 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.668735 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:56Z","lastTransitionTime":"2025-12-02T01:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.772547 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.772620 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.772644 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.772675 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.772700 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:56Z","lastTransitionTime":"2025-12-02T01:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.875607 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.875668 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.875685 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.875709 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.875726 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:56Z","lastTransitionTime":"2025-12-02T01:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.978722 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.978807 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.978825 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.978849 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:56 crc kubenswrapper[4884]: I1202 01:37:56.978866 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:56Z","lastTransitionTime":"2025-12-02T01:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.082525 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.082627 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.082645 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.082702 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.082719 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:57Z","lastTransitionTime":"2025-12-02T01:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.185984 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.186062 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.186087 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.186117 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.186139 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:57Z","lastTransitionTime":"2025-12-02T01:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.288862 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.288938 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.288963 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.288988 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.289008 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:57Z","lastTransitionTime":"2025-12-02T01:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.391599 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.391649 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.391664 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.391685 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.391700 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:57Z","lastTransitionTime":"2025-12-02T01:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.494368 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.494441 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.494463 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.494498 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.494521 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:57Z","lastTransitionTime":"2025-12-02T01:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.597250 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.597330 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.597355 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.597421 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.597446 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:57Z","lastTransitionTime":"2025-12-02T01:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.700813 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.700877 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.700893 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.700915 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.700935 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:57Z","lastTransitionTime":"2025-12-02T01:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.804445 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.804504 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.804521 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.804544 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.804562 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:57Z","lastTransitionTime":"2025-12-02T01:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.907911 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.907979 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.907997 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.908021 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:57 crc kubenswrapper[4884]: I1202 01:37:57.908040 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:57Z","lastTransitionTime":"2025-12-02T01:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.011250 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.011320 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.011344 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.011375 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.011399 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:58Z","lastTransitionTime":"2025-12-02T01:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.114181 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.114495 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.114508 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.114524 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.114537 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:58Z","lastTransitionTime":"2025-12-02T01:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.218347 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.218417 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.218438 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.218468 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.218491 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:58Z","lastTransitionTime":"2025-12-02T01:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.321532 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.321576 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.321592 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.321614 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.321629 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:58Z","lastTransitionTime":"2025-12-02T01:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.423850 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.423911 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.423928 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.423952 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.423968 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:58Z","lastTransitionTime":"2025-12-02T01:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.526759 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.526793 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.526805 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.526822 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.526833 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:58Z","lastTransitionTime":"2025-12-02T01:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.613804 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.613880 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.613952 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.613806 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:37:58 crc kubenswrapper[4884]: E1202 01:37:58.613995 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:37:58 crc kubenswrapper[4884]: E1202 01:37:58.614163 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:37:58 crc kubenswrapper[4884]: E1202 01:37:58.614396 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:37:58 crc kubenswrapper[4884]: E1202 01:37:58.614506 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.629239 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.629300 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.629316 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.629340 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.629361 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:58Z","lastTransitionTime":"2025-12-02T01:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.650996 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.651063 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.651087 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.651120 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.651146 4884 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-02T01:37:58Z","lastTransitionTime":"2025-12-02T01:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.727803 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj"] Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.728439 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.731498 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.731661 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.732830 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.734370 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.796432 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-fdrt4" podStartSLOduration=76.796411635 podStartE2EDuration="1m16.796411635s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:37:58.79535834 +0000 UTC m=+95.471195284" watchObservedRunningTime="2025-12-02 01:37:58.796411635 +0000 UTC m=+95.472248559" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.834142 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-ww4r4" podStartSLOduration=75.834107492 podStartE2EDuration="1m15.834107492s" podCreationTimestamp="2025-12-02 01:36:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:37:58.833665511 +0000 UTC m=+95.509502455" watchObservedRunningTime="2025-12-02 01:37:58.834107492 +0000 UTC m=+95.509944416" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.853313 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/795fded5-bfad-4c28-91f5-5bd0452eec56-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-pd9qj\" (UID: \"795fded5-bfad-4c28-91f5-5bd0452eec56\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.853415 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/795fded5-bfad-4c28-91f5-5bd0452eec56-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-pd9qj\" (UID: \"795fded5-bfad-4c28-91f5-5bd0452eec56\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.853503 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/795fded5-bfad-4c28-91f5-5bd0452eec56-service-ca\") pod \"cluster-version-operator-5c965bbfc6-pd9qj\" (UID: \"795fded5-bfad-4c28-91f5-5bd0452eec56\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.853592 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/795fded5-bfad-4c28-91f5-5bd0452eec56-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-pd9qj\" (UID: \"795fded5-bfad-4c28-91f5-5bd0452eec56\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.853720 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/795fded5-bfad-4c28-91f5-5bd0452eec56-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-pd9qj\" (UID: \"795fded5-bfad-4c28-91f5-5bd0452eec56\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.856487 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=5.856462565 podStartE2EDuration="5.856462565s" podCreationTimestamp="2025-12-02 01:37:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:37:58.853114854 +0000 UTC m=+95.528951798" watchObservedRunningTime="2025-12-02 01:37:58.856462565 +0000 UTC m=+95.532299479" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.955351 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/795fded5-bfad-4c28-91f5-5bd0452eec56-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-pd9qj\" (UID: \"795fded5-bfad-4c28-91f5-5bd0452eec56\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.955502 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/795fded5-bfad-4c28-91f5-5bd0452eec56-service-ca\") pod \"cluster-version-operator-5c965bbfc6-pd9qj\" (UID: \"795fded5-bfad-4c28-91f5-5bd0452eec56\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.955551 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/795fded5-bfad-4c28-91f5-5bd0452eec56-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-pd9qj\" (UID: \"795fded5-bfad-4c28-91f5-5bd0452eec56\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.955606 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/795fded5-bfad-4c28-91f5-5bd0452eec56-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-pd9qj\" (UID: \"795fded5-bfad-4c28-91f5-5bd0452eec56\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.955719 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/795fded5-bfad-4c28-91f5-5bd0452eec56-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-pd9qj\" (UID: \"795fded5-bfad-4c28-91f5-5bd0452eec56\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.955898 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/795fded5-bfad-4c28-91f5-5bd0452eec56-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-pd9qj\" (UID: \"795fded5-bfad-4c28-91f5-5bd0452eec56\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.956030 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/795fded5-bfad-4c28-91f5-5bd0452eec56-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-pd9qj\" (UID: \"795fded5-bfad-4c28-91f5-5bd0452eec56\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.957439 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/795fded5-bfad-4c28-91f5-5bd0452eec56-service-ca\") pod \"cluster-version-operator-5c965bbfc6-pd9qj\" (UID: \"795fded5-bfad-4c28-91f5-5bd0452eec56\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.963484 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/795fded5-bfad-4c28-91f5-5bd0452eec56-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-pd9qj\" (UID: \"795fded5-bfad-4c28-91f5-5bd0452eec56\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:58 crc kubenswrapper[4884]: I1202 01:37:58.984976 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/795fded5-bfad-4c28-91f5-5bd0452eec56-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-pd9qj\" (UID: \"795fded5-bfad-4c28-91f5-5bd0452eec56\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:59 crc kubenswrapper[4884]: I1202 01:37:59.056775 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" Dec 02 01:37:59 crc kubenswrapper[4884]: W1202 01:37:59.078291 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod795fded5_bfad_4c28_91f5_5bd0452eec56.slice/crio-272a5dc7a96df5445561999652e1cc894fdfc422d197561720bd3a7be2b28b41 WatchSource:0}: Error finding container 272a5dc7a96df5445561999652e1cc894fdfc422d197561720bd3a7be2b28b41: Status 404 returned error can't find the container with id 272a5dc7a96df5445561999652e1cc894fdfc422d197561720bd3a7be2b28b41 Dec 02 01:37:59 crc kubenswrapper[4884]: I1202 01:37:59.429427 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" event={"ID":"795fded5-bfad-4c28-91f5-5bd0452eec56","Type":"ContainerStarted","Data":"457480adb8487f6ff7f7071543aca248952776f84292f0eaf72ce755d0ba8546"} Dec 02 01:37:59 crc kubenswrapper[4884]: I1202 01:37:59.429525 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" event={"ID":"795fded5-bfad-4c28-91f5-5bd0452eec56","Type":"ContainerStarted","Data":"272a5dc7a96df5445561999652e1cc894fdfc422d197561720bd3a7be2b28b41"} Dec 02 01:37:59 crc kubenswrapper[4884]: I1202 01:37:59.448785 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pd9qj" podStartSLOduration=77.448736308 podStartE2EDuration="1m17.448736308s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:37:59.446944354 +0000 UTC m=+96.122781268" watchObservedRunningTime="2025-12-02 01:37:59.448736308 +0000 UTC m=+96.124573222" Dec 02 01:38:00 crc kubenswrapper[4884]: I1202 01:38:00.613208 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:00 crc kubenswrapper[4884]: I1202 01:38:00.613227 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:00 crc kubenswrapper[4884]: I1202 01:38:00.613345 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:00 crc kubenswrapper[4884]: I1202 01:38:00.613399 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:00 crc kubenswrapper[4884]: E1202 01:38:00.613507 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:00 crc kubenswrapper[4884]: E1202 01:38:00.613612 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:00 crc kubenswrapper[4884]: E1202 01:38:00.613674 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:00 crc kubenswrapper[4884]: E1202 01:38:00.613899 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:00 crc kubenswrapper[4884]: I1202 01:38:00.773214 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs\") pod \"network-metrics-daemon-whh6t\" (UID: \"7a6a67e3-3b98-45df-9573-6efee410a90c\") " pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:00 crc kubenswrapper[4884]: E1202 01:38:00.773450 4884 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:38:00 crc kubenswrapper[4884]: E1202 01:38:00.773590 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs podName:7a6a67e3-3b98-45df-9573-6efee410a90c nodeName:}" failed. No retries permitted until 2025-12-02 01:39:04.773559052 +0000 UTC m=+161.449395976 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs") pod "network-metrics-daemon-whh6t" (UID: "7a6a67e3-3b98-45df-9573-6efee410a90c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 02 01:38:02 crc kubenswrapper[4884]: I1202 01:38:02.613871 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:02 crc kubenswrapper[4884]: I1202 01:38:02.613892 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:02 crc kubenswrapper[4884]: E1202 01:38:02.614334 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:02 crc kubenswrapper[4884]: I1202 01:38:02.614008 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:02 crc kubenswrapper[4884]: I1202 01:38:02.613965 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:02 crc kubenswrapper[4884]: E1202 01:38:02.614489 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:02 crc kubenswrapper[4884]: E1202 01:38:02.614567 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:02 crc kubenswrapper[4884]: E1202 01:38:02.614644 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:04 crc kubenswrapper[4884]: I1202 01:38:04.614115 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:04 crc kubenswrapper[4884]: I1202 01:38:04.614189 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:04 crc kubenswrapper[4884]: E1202 01:38:04.614313 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:04 crc kubenswrapper[4884]: E1202 01:38:04.615043 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:04 crc kubenswrapper[4884]: I1202 01:38:04.614150 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:04 crc kubenswrapper[4884]: E1202 01:38:04.615211 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:04 crc kubenswrapper[4884]: I1202 01:38:04.615520 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:04 crc kubenswrapper[4884]: E1202 01:38:04.615704 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:06 crc kubenswrapper[4884]: I1202 01:38:06.613680 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:06 crc kubenswrapper[4884]: I1202 01:38:06.613855 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:06 crc kubenswrapper[4884]: I1202 01:38:06.613921 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:06 crc kubenswrapper[4884]: E1202 01:38:06.614000 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:06 crc kubenswrapper[4884]: I1202 01:38:06.613729 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:06 crc kubenswrapper[4884]: E1202 01:38:06.614469 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:06 crc kubenswrapper[4884]: E1202 01:38:06.614942 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:06 crc kubenswrapper[4884]: E1202 01:38:06.615152 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:07 crc kubenswrapper[4884]: I1202 01:38:07.615323 4884 scope.go:117] "RemoveContainer" containerID="5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9" Dec 02 01:38:07 crc kubenswrapper[4884]: E1202 01:38:07.615571 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-c74x8_openshift-ovn-kubernetes(d8c50c39-2888-444e-bf55-49b091559978)\"" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" Dec 02 01:38:08 crc kubenswrapper[4884]: I1202 01:38:08.613954 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:08 crc kubenswrapper[4884]: I1202 01:38:08.613967 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:08 crc kubenswrapper[4884]: I1202 01:38:08.614000 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:08 crc kubenswrapper[4884]: E1202 01:38:08.614595 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:08 crc kubenswrapper[4884]: E1202 01:38:08.614425 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:08 crc kubenswrapper[4884]: I1202 01:38:08.614038 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:08 crc kubenswrapper[4884]: E1202 01:38:08.614790 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:08 crc kubenswrapper[4884]: E1202 01:38:08.615063 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:10 crc kubenswrapper[4884]: I1202 01:38:10.613999 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:10 crc kubenswrapper[4884]: I1202 01:38:10.614013 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:10 crc kubenswrapper[4884]: I1202 01:38:10.614293 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:10 crc kubenswrapper[4884]: E1202 01:38:10.614234 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:10 crc kubenswrapper[4884]: I1202 01:38:10.614025 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:10 crc kubenswrapper[4884]: E1202 01:38:10.614854 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:10 crc kubenswrapper[4884]: E1202 01:38:10.615016 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:10 crc kubenswrapper[4884]: E1202 01:38:10.615189 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:12 crc kubenswrapper[4884]: I1202 01:38:12.613920 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:12 crc kubenswrapper[4884]: I1202 01:38:12.614023 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:12 crc kubenswrapper[4884]: I1202 01:38:12.613954 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:12 crc kubenswrapper[4884]: I1202 01:38:12.613953 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:12 crc kubenswrapper[4884]: E1202 01:38:12.614214 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:12 crc kubenswrapper[4884]: E1202 01:38:12.614521 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:12 crc kubenswrapper[4884]: E1202 01:38:12.614642 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:12 crc kubenswrapper[4884]: E1202 01:38:12.614379 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:14 crc kubenswrapper[4884]: I1202 01:38:14.613658 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:14 crc kubenswrapper[4884]: I1202 01:38:14.613708 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:14 crc kubenswrapper[4884]: I1202 01:38:14.613889 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:14 crc kubenswrapper[4884]: E1202 01:38:14.614092 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:14 crc kubenswrapper[4884]: I1202 01:38:14.614161 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:14 crc kubenswrapper[4884]: E1202 01:38:14.614335 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:14 crc kubenswrapper[4884]: E1202 01:38:14.614462 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:14 crc kubenswrapper[4884]: E1202 01:38:14.614565 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:16 crc kubenswrapper[4884]: I1202 01:38:16.487167 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6l592_791af0d6-a201-430f-b09d-02c24a4b7503/kube-multus/1.log" Dec 02 01:38:16 crc kubenswrapper[4884]: I1202 01:38:16.488572 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6l592_791af0d6-a201-430f-b09d-02c24a4b7503/kube-multus/0.log" Dec 02 01:38:16 crc kubenswrapper[4884]: I1202 01:38:16.488636 4884 generic.go:334] "Generic (PLEG): container finished" podID="791af0d6-a201-430f-b09d-02c24a4b7503" containerID="b3581bf807ed4bcf2bc452f0ad29348534c8c44bdc2d6a24614c2ea43c6d63a3" exitCode=1 Dec 02 01:38:16 crc kubenswrapper[4884]: I1202 01:38:16.488687 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6l592" event={"ID":"791af0d6-a201-430f-b09d-02c24a4b7503","Type":"ContainerDied","Data":"b3581bf807ed4bcf2bc452f0ad29348534c8c44bdc2d6a24614c2ea43c6d63a3"} Dec 02 01:38:16 crc kubenswrapper[4884]: I1202 01:38:16.488781 4884 scope.go:117] "RemoveContainer" containerID="4c7f199a5d7a51890d7a47397c52f645467bf20483f1b4e8af0abcda6b850a95" Dec 02 01:38:16 crc kubenswrapper[4884]: I1202 01:38:16.489376 4884 scope.go:117] "RemoveContainer" containerID="b3581bf807ed4bcf2bc452f0ad29348534c8c44bdc2d6a24614c2ea43c6d63a3" Dec 02 01:38:16 crc kubenswrapper[4884]: E1202 01:38:16.489865 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-6l592_openshift-multus(791af0d6-a201-430f-b09d-02c24a4b7503)\"" pod="openshift-multus/multus-6l592" podUID="791af0d6-a201-430f-b09d-02c24a4b7503" Dec 02 01:38:16 crc kubenswrapper[4884]: I1202 01:38:16.613793 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:16 crc kubenswrapper[4884]: I1202 01:38:16.613843 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:16 crc kubenswrapper[4884]: I1202 01:38:16.613858 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:16 crc kubenswrapper[4884]: I1202 01:38:16.613793 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:16 crc kubenswrapper[4884]: E1202 01:38:16.613975 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:16 crc kubenswrapper[4884]: E1202 01:38:16.614063 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:16 crc kubenswrapper[4884]: E1202 01:38:16.614160 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:16 crc kubenswrapper[4884]: E1202 01:38:16.614320 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:17 crc kubenswrapper[4884]: I1202 01:38:17.494871 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6l592_791af0d6-a201-430f-b09d-02c24a4b7503/kube-multus/1.log" Dec 02 01:38:18 crc kubenswrapper[4884]: I1202 01:38:18.613084 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:18 crc kubenswrapper[4884]: I1202 01:38:18.613410 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:18 crc kubenswrapper[4884]: E1202 01:38:18.613603 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:18 crc kubenswrapper[4884]: I1202 01:38:18.613703 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:18 crc kubenswrapper[4884]: E1202 01:38:18.613855 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:18 crc kubenswrapper[4884]: I1202 01:38:18.613926 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:18 crc kubenswrapper[4884]: E1202 01:38:18.613994 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:18 crc kubenswrapper[4884]: E1202 01:38:18.614119 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:20 crc kubenswrapper[4884]: I1202 01:38:20.614084 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:20 crc kubenswrapper[4884]: I1202 01:38:20.614137 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:20 crc kubenswrapper[4884]: I1202 01:38:20.614170 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:20 crc kubenswrapper[4884]: I1202 01:38:20.614308 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:20 crc kubenswrapper[4884]: E1202 01:38:20.614645 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:20 crc kubenswrapper[4884]: E1202 01:38:20.614830 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:20 crc kubenswrapper[4884]: E1202 01:38:20.615583 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:20 crc kubenswrapper[4884]: E1202 01:38:20.615782 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:20 crc kubenswrapper[4884]: I1202 01:38:20.616129 4884 scope.go:117] "RemoveContainer" containerID="5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9" Dec 02 01:38:21 crc kubenswrapper[4884]: I1202 01:38:21.513827 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/3.log" Dec 02 01:38:21 crc kubenswrapper[4884]: I1202 01:38:21.517602 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerStarted","Data":"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2"} Dec 02 01:38:21 crc kubenswrapper[4884]: I1202 01:38:21.518009 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:38:21 crc kubenswrapper[4884]: I1202 01:38:21.561560 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podStartSLOduration=99.56153498 podStartE2EDuration="1m39.56153498s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:21.560947896 +0000 UTC m=+118.236784810" watchObservedRunningTime="2025-12-02 01:38:21.56153498 +0000 UTC m=+118.237371904" Dec 02 01:38:21 crc kubenswrapper[4884]: I1202 01:38:21.662402 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-whh6t"] Dec 02 01:38:21 crc kubenswrapper[4884]: I1202 01:38:21.662562 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:21 crc kubenswrapper[4884]: E1202 01:38:21.662697 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:22 crc kubenswrapper[4884]: I1202 01:38:22.613903 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:22 crc kubenswrapper[4884]: I1202 01:38:22.613954 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:22 crc kubenswrapper[4884]: E1202 01:38:22.614455 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:22 crc kubenswrapper[4884]: I1202 01:38:22.613970 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:22 crc kubenswrapper[4884]: E1202 01:38:22.614547 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:22 crc kubenswrapper[4884]: E1202 01:38:22.614689 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:23 crc kubenswrapper[4884]: E1202 01:38:23.609629 4884 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 02 01:38:23 crc kubenswrapper[4884]: I1202 01:38:23.614087 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:23 crc kubenswrapper[4884]: E1202 01:38:23.615916 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:23 crc kubenswrapper[4884]: E1202 01:38:23.957587 4884 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 02 01:38:24 crc kubenswrapper[4884]: I1202 01:38:24.613539 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:24 crc kubenswrapper[4884]: I1202 01:38:24.613596 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:24 crc kubenswrapper[4884]: I1202 01:38:24.613615 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:24 crc kubenswrapper[4884]: E1202 01:38:24.613710 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:24 crc kubenswrapper[4884]: E1202 01:38:24.614184 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:24 crc kubenswrapper[4884]: E1202 01:38:24.614488 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:25 crc kubenswrapper[4884]: I1202 01:38:25.613182 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:25 crc kubenswrapper[4884]: E1202 01:38:25.613818 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:26 crc kubenswrapper[4884]: I1202 01:38:26.614100 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:26 crc kubenswrapper[4884]: I1202 01:38:26.614139 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:26 crc kubenswrapper[4884]: I1202 01:38:26.614189 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:26 crc kubenswrapper[4884]: E1202 01:38:26.614255 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:26 crc kubenswrapper[4884]: E1202 01:38:26.614406 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:26 crc kubenswrapper[4884]: E1202 01:38:26.614518 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:27 crc kubenswrapper[4884]: I1202 01:38:27.613450 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:27 crc kubenswrapper[4884]: E1202 01:38:27.613652 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:27 crc kubenswrapper[4884]: I1202 01:38:27.614363 4884 scope.go:117] "RemoveContainer" containerID="b3581bf807ed4bcf2bc452f0ad29348534c8c44bdc2d6a24614c2ea43c6d63a3" Dec 02 01:38:28 crc kubenswrapper[4884]: I1202 01:38:28.544932 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6l592_791af0d6-a201-430f-b09d-02c24a4b7503/kube-multus/1.log" Dec 02 01:38:28 crc kubenswrapper[4884]: I1202 01:38:28.545692 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6l592" event={"ID":"791af0d6-a201-430f-b09d-02c24a4b7503","Type":"ContainerStarted","Data":"e1c1a0f1f733644cf5808f88a04f9b1c235bcc20b998175fd39ea33a4a38e6de"} Dec 02 01:38:28 crc kubenswrapper[4884]: I1202 01:38:28.613598 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:28 crc kubenswrapper[4884]: I1202 01:38:28.613660 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:28 crc kubenswrapper[4884]: I1202 01:38:28.614128 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:28 crc kubenswrapper[4884]: E1202 01:38:28.614269 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:28 crc kubenswrapper[4884]: E1202 01:38:28.614471 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:28 crc kubenswrapper[4884]: E1202 01:38:28.614664 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:28 crc kubenswrapper[4884]: E1202 01:38:28.959336 4884 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 02 01:38:29 crc kubenswrapper[4884]: I1202 01:38:29.613770 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:29 crc kubenswrapper[4884]: E1202 01:38:29.613949 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:30 crc kubenswrapper[4884]: I1202 01:38:30.613509 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:30 crc kubenswrapper[4884]: I1202 01:38:30.613577 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:30 crc kubenswrapper[4884]: E1202 01:38:30.613683 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:30 crc kubenswrapper[4884]: I1202 01:38:30.613545 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:30 crc kubenswrapper[4884]: E1202 01:38:30.613911 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:30 crc kubenswrapper[4884]: E1202 01:38:30.614017 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:31 crc kubenswrapper[4884]: I1202 01:38:31.613435 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:31 crc kubenswrapper[4884]: E1202 01:38:31.614159 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:32 crc kubenswrapper[4884]: I1202 01:38:32.613691 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:32 crc kubenswrapper[4884]: I1202 01:38:32.613730 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:32 crc kubenswrapper[4884]: I1202 01:38:32.613796 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:32 crc kubenswrapper[4884]: E1202 01:38:32.613835 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 02 01:38:32 crc kubenswrapper[4884]: E1202 01:38:32.613960 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 02 01:38:32 crc kubenswrapper[4884]: E1202 01:38:32.614009 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 02 01:38:33 crc kubenswrapper[4884]: I1202 01:38:33.613700 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:33 crc kubenswrapper[4884]: E1202 01:38:33.615109 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-whh6t" podUID="7a6a67e3-3b98-45df-9573-6efee410a90c" Dec 02 01:38:34 crc kubenswrapper[4884]: I1202 01:38:34.613675 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:34 crc kubenswrapper[4884]: I1202 01:38:34.613683 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:34 crc kubenswrapper[4884]: I1202 01:38:34.613692 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:34 crc kubenswrapper[4884]: I1202 01:38:34.617377 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 02 01:38:34 crc kubenswrapper[4884]: I1202 01:38:34.617387 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 02 01:38:34 crc kubenswrapper[4884]: I1202 01:38:34.617398 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 02 01:38:34 crc kubenswrapper[4884]: I1202 01:38:34.617535 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 02 01:38:35 crc kubenswrapper[4884]: I1202 01:38:35.613920 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:38:35 crc kubenswrapper[4884]: I1202 01:38:35.617184 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 02 01:38:35 crc kubenswrapper[4884]: I1202 01:38:35.617492 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.509489 4884 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.561569 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.562442 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.563044 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.563827 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.564786 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4htsc"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.565597 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.566406 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.567125 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.568076 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mtv5w"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.568720 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.569789 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.570542 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.573870 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.575068 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.575150 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.575153 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.575257 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.579153 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sfn66"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.588884 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.589463 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.602595 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n7tvs"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.605192 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.605357 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.607178 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.607618 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vwnln"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.608954 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n7tvs" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.635009 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.635876 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.636053 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.636223 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.636405 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.636693 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.641682 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.641805 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.641858 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.641913 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.641933 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642022 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642243 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642299 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642315 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642509 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642572 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642611 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642632 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642697 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642704 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642578 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642517 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642517 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642813 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7v5b\" (UniqueName: \"kubernetes.io/projected/041d3d0d-70e0-4249-a7e2-63958513ba13-kube-api-access-q7v5b\") pod \"route-controller-manager-6576b87f9c-lqhjg\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642867 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x9ws\" (UniqueName: \"kubernetes.io/projected/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-kube-api-access-4x9ws\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642922 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.642979 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643092 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0d3cbb8-cbf0-4799-b40f-f97beb16a022-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6qvr7\" (UID: \"f0d3cbb8-cbf0-4799-b40f-f97beb16a022\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643128 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smp4d\" (UniqueName: \"kubernetes.io/projected/2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4-kube-api-access-smp4d\") pod \"machine-api-operator-5694c8668f-4htsc\" (UID: \"2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643156 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8e078cc-b142-4ccb-b3e1-5c899a354113-config\") pod \"authentication-operator-69f744f599-mtv5w\" (UID: \"d8e078cc-b142-4ccb-b3e1-5c899a354113\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643200 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b18452c-12ba-4d43-b858-4608001e76d7-config\") pod \"machine-approver-56656f9798-jhnjt\" (UID: \"6b18452c-12ba-4d43-b858-4608001e76d7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643211 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643230 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8e078cc-b142-4ccb-b3e1-5c899a354113-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mtv5w\" (UID: \"d8e078cc-b142-4ccb-b3e1-5c899a354113\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643266 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4htsc\" (UID: \"2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643304 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4-config\") pod \"machine-api-operator-5694c8668f-4htsc\" (UID: \"2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643333 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4-images\") pod \"machine-api-operator-5694c8668f-4htsc\" (UID: \"2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643372 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/041d3d0d-70e0-4249-a7e2-63958513ba13-serving-cert\") pod \"route-controller-manager-6576b87f9c-lqhjg\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643397 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0d3cbb8-cbf0-4799-b40f-f97beb16a022-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6qvr7\" (UID: \"f0d3cbb8-cbf0-4799-b40f-f97beb16a022\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643428 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-audit-policies\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643447 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643450 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rht9g\" (UniqueName: \"kubernetes.io/projected/d8e078cc-b142-4ccb-b3e1-5c899a354113-kube-api-access-rht9g\") pod \"authentication-operator-69f744f599-mtv5w\" (UID: \"d8e078cc-b142-4ccb-b3e1-5c899a354113\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643478 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-serving-cert\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643504 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6b18452c-12ba-4d43-b858-4608001e76d7-machine-approver-tls\") pod \"machine-approver-56656f9798-jhnjt\" (UID: \"6b18452c-12ba-4d43-b858-4608001e76d7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643529 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8e078cc-b142-4ccb-b3e1-5c899a354113-service-ca-bundle\") pod \"authentication-operator-69f744f599-mtv5w\" (UID: \"d8e078cc-b142-4ccb-b3e1-5c899a354113\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643564 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/041d3d0d-70e0-4249-a7e2-63958513ba13-config\") pod \"route-controller-manager-6576b87f9c-lqhjg\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643586 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e078cc-b142-4ccb-b3e1-5c899a354113-serving-cert\") pod \"authentication-operator-69f744f599-mtv5w\" (UID: \"d8e078cc-b142-4ccb-b3e1-5c899a354113\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643609 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnm4m\" (UniqueName: \"kubernetes.io/projected/f0d3cbb8-cbf0-4799-b40f-f97beb16a022-kube-api-access-lnm4m\") pod \"openshift-apiserver-operator-796bbdcf4f-6qvr7\" (UID: \"f0d3cbb8-cbf0-4799-b40f-f97beb16a022\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643630 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcp99\" (UniqueName: \"kubernetes.io/projected/6b18452c-12ba-4d43-b858-4608001e76d7-kube-api-access-hcp99\") pod \"machine-approver-56656f9798-jhnjt\" (UID: \"6b18452c-12ba-4d43-b858-4608001e76d7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643660 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-encryption-config\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643690 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6b18452c-12ba-4d43-b858-4608001e76d7-auth-proxy-config\") pod \"machine-approver-56656f9798-jhnjt\" (UID: \"6b18452c-12ba-4d43-b858-4608001e76d7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643812 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.643975 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.644116 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.644260 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.644395 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.644526 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.647793 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-audit-dir\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.647841 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.647885 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-etcd-client\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.647932 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/041d3d0d-70e0-4249-a7e2-63958513ba13-client-ca\") pod \"route-controller-manager-6576b87f9c-lqhjg\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.648962 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.649692 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.655544 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.656107 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.656229 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.656390 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.656675 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.657680 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.657791 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.657839 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.657845 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.658444 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.658571 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.659826 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-mh86q"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.660205 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-pmdpb"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.660452 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-tk4v5"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.660834 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n757x"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.661302 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.661666 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.661916 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-pmdpb" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.662099 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-tk4v5" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.670998 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.671193 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.671300 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.671417 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.673283 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.673612 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.673823 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wpxzm"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.674414 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.676357 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.676723 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.678124 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.678649 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.682060 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.682531 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.682767 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.682895 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.683016 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.683357 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.683563 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.683724 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.684341 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.684704 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.684999 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.685890 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.686187 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.686231 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.686309 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.695964 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.698940 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.699487 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.700677 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.700825 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-pcf5t"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.702235 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.702287 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.705932 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.706311 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.706341 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.706945 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.707083 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.707174 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.732194 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.732328 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.732550 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.732772 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.735598 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-664p9"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.735899 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.736126 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.736282 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cmk67"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.736333 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.736457 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.736570 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.736684 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.736727 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.736908 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.736955 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.737222 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.738264 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.738605 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.740455 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.741381 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.741821 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.741834 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.748432 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.749137 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750052 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0d3cbb8-cbf0-4799-b40f-f97beb16a022-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6qvr7\" (UID: \"f0d3cbb8-cbf0-4799-b40f-f97beb16a022\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750098 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ef99886-22f3-4634-9715-0dd2ee3c53a7-serving-cert\") pod \"console-operator-58897d9998-vwnln\" (UID: \"5ef99886-22f3-4634-9715-0dd2ee3c53a7\") " pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750131 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-config\") pod \"controller-manager-879f6c89f-n757x\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750154 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-audit-policies\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750178 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rht9g\" (UniqueName: \"kubernetes.io/projected/d8e078cc-b142-4ccb-b3e1-5c899a354113-kube-api-access-rht9g\") pod \"authentication-operator-69f744f599-mtv5w\" (UID: \"d8e078cc-b142-4ccb-b3e1-5c899a354113\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750202 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750301 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96x9n\" (UniqueName: \"kubernetes.io/projected/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-kube-api-access-96x9n\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750350 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ef99886-22f3-4634-9715-0dd2ee3c53a7-config\") pod \"console-operator-58897d9998-vwnln\" (UID: \"5ef99886-22f3-4634-9715-0dd2ee3c53a7\") " pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750386 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-serving-cert\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750405 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj5lw\" (UniqueName: \"kubernetes.io/projected/3b0da047-6c65-43c9-b8c9-5fef3c6a7929-kube-api-access-rj5lw\") pod \"openshift-config-operator-7777fb866f-bbcm8\" (UID: \"3b0da047-6c65-43c9-b8c9-5fef3c6a7929\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750429 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750446 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750462 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5a019c2c-939e-4d88-b662-79dd417dce18-etcd-client\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750478 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccc071ec-9724-41f2-a654-625aae8bd840-serving-cert\") pod \"controller-manager-879f6c89f-n757x\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750496 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6b18452c-12ba-4d43-b858-4608001e76d7-machine-approver-tls\") pod \"machine-approver-56656f9798-jhnjt\" (UID: \"6b18452c-12ba-4d43-b858-4608001e76d7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750527 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8e078cc-b142-4ccb-b3e1-5c899a354113-service-ca-bundle\") pod \"authentication-operator-69f744f599-mtv5w\" (UID: \"d8e078cc-b142-4ccb-b3e1-5c899a354113\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750063 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750145 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750796 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-audit-policies\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.750332 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751044 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751242 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751293 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751344 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/041d3d0d-70e0-4249-a7e2-63958513ba13-config\") pod \"route-controller-manager-6576b87f9c-lqhjg\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751375 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e078cc-b142-4ccb-b3e1-5c899a354113-serving-cert\") pod \"authentication-operator-69f744f599-mtv5w\" (UID: \"d8e078cc-b142-4ccb-b3e1-5c899a354113\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751399 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3b0da047-6c65-43c9-b8c9-5fef3c6a7929-available-featuregates\") pod \"openshift-config-operator-7777fb866f-bbcm8\" (UID: \"3b0da047-6c65-43c9-b8c9-5fef3c6a7929\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751425 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnm4m\" (UniqueName: \"kubernetes.io/projected/f0d3cbb8-cbf0-4799-b40f-f97beb16a022-kube-api-access-lnm4m\") pod \"openshift-apiserver-operator-796bbdcf4f-6qvr7\" (UID: \"f0d3cbb8-cbf0-4799-b40f-f97beb16a022\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751451 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcp99\" (UniqueName: \"kubernetes.io/projected/6b18452c-12ba-4d43-b858-4608001e76d7-kube-api-access-hcp99\") pod \"machine-approver-56656f9798-jhnjt\" (UID: \"6b18452c-12ba-4d43-b858-4608001e76d7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751555 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751584 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzv9t\" (UniqueName: \"kubernetes.io/projected/5ef99886-22f3-4634-9715-0dd2ee3c53a7-kube-api-access-mzv9t\") pod \"console-operator-58897d9998-vwnln\" (UID: \"5ef99886-22f3-4634-9715-0dd2ee3c53a7\") " pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751612 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3e56bc10-6d63-4b7b-a740-b449267b3c20-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fjz74\" (UID: \"3e56bc10-6d63-4b7b-a740-b449267b3c20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751637 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b0da047-6c65-43c9-b8c9-5fef3c6a7929-serving-cert\") pod \"openshift-config-operator-7777fb866f-bbcm8\" (UID: \"3b0da047-6c65-43c9-b8c9-5fef3c6a7929\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751672 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5a019c2c-939e-4d88-b662-79dd417dce18-image-import-ca\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751697 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-oauth-serving-cert\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751738 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-encryption-config\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751815 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751872 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e56bc10-6d63-4b7b-a740-b449267b3c20-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fjz74\" (UID: \"3e56bc10-6d63-4b7b-a740-b449267b3c20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751906 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-audit-dir\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751934 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6b18452c-12ba-4d43-b858-4608001e76d7-auth-proxy-config\") pod \"machine-approver-56656f9798-jhnjt\" (UID: \"6b18452c-12ba-4d43-b858-4608001e76d7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751963 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-service-ca\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.751987 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ef99886-22f3-4634-9715-0dd2ee3c53a7-trusted-ca\") pod \"console-operator-58897d9998-vwnln\" (UID: \"5ef99886-22f3-4634-9715-0dd2ee3c53a7\") " pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752013 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-trusted-ca-bundle\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752041 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752066 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752090 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z495x\" (UniqueName: \"kubernetes.io/projected/0b33acfc-c22d-4d1f-a70f-410ed90474bb-kube-api-access-z495x\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752113 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5a019c2c-939e-4d88-b662-79dd417dce18-node-pullsecrets\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752147 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c82x\" (UniqueName: \"kubernetes.io/projected/3e56bc10-6d63-4b7b-a740-b449267b3c20-kube-api-access-4c82x\") pod \"cluster-image-registry-operator-dc59b4c8b-fjz74\" (UID: \"3e56bc10-6d63-4b7b-a740-b449267b3c20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752175 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-etcd-client\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752196 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752226 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752248 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-serving-cert\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752273 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85a7c94e-bc28-4d2e-ad77-7f776252b482-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wthrm\" (UID: \"85a7c94e-bc28-4d2e-ad77-7f776252b482\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752299 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/15377bb0-a5a9-4a17-bd32-32795886deac-metrics-tls\") pod \"dns-operator-744455d44c-tk4v5\" (UID: \"15377bb0-a5a9-4a17-bd32-32795886deac\") " pod="openshift-dns-operator/dns-operator-744455d44c-tk4v5" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752324 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/041d3d0d-70e0-4249-a7e2-63958513ba13-client-ca\") pod \"route-controller-manager-6576b87f9c-lqhjg\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752350 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2m45\" (UniqueName: \"kubernetes.io/projected/5a019c2c-939e-4d88-b662-79dd417dce18-kube-api-access-w2m45\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752373 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-client-ca\") pod \"controller-manager-879f6c89f-n757x\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752400 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e56bc10-6d63-4b7b-a740-b449267b3c20-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fjz74\" (UID: \"3e56bc10-6d63-4b7b-a740-b449267b3c20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752424 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0b33acfc-c22d-4d1f-a70f-410ed90474bb-audit-dir\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752464 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7v5b\" (UniqueName: \"kubernetes.io/projected/041d3d0d-70e0-4249-a7e2-63958513ba13-kube-api-access-q7v5b\") pod \"route-controller-manager-6576b87f9c-lqhjg\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752491 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5a019c2c-939e-4d88-b662-79dd417dce18-etcd-serving-ca\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752828 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a019c2c-939e-4d88-b662-79dd417dce18-serving-cert\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.752959 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a019c2c-939e-4d88-b662-79dd417dce18-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753009 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753034 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x9ws\" (UniqueName: \"kubernetes.io/projected/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-kube-api-access-4x9ws\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753059 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-oauth-config\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753096 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a019c2c-939e-4d88-b662-79dd417dce18-config\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753119 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/d62437d3-3b65-4029-9f01-f6a696e40895-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-n7tvs\" (UID: \"d62437d3-3b65-4029-9f01-f6a696e40895\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n7tvs" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753146 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0d3cbb8-cbf0-4799-b40f-f97beb16a022-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6qvr7\" (UID: \"f0d3cbb8-cbf0-4799-b40f-f97beb16a022\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753170 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753196 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc6ht\" (UniqueName: \"kubernetes.io/projected/85a7c94e-bc28-4d2e-ad77-7f776252b482-kube-api-access-tc6ht\") pod \"openshift-controller-manager-operator-756b6f6bc6-wthrm\" (UID: \"85a7c94e-bc28-4d2e-ad77-7f776252b482\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753225 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b18452c-12ba-4d43-b858-4608001e76d7-config\") pod \"machine-approver-56656f9798-jhnjt\" (UID: \"6b18452c-12ba-4d43-b858-4608001e76d7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753249 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smp4d\" (UniqueName: \"kubernetes.io/projected/2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4-kube-api-access-smp4d\") pod \"machine-api-operator-5694c8668f-4htsc\" (UID: \"2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753275 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8e078cc-b142-4ccb-b3e1-5c899a354113-config\") pod \"authentication-operator-69f744f599-mtv5w\" (UID: \"d8e078cc-b142-4ccb-b3e1-5c899a354113\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753298 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5a019c2c-939e-4d88-b662-79dd417dce18-audit-dir\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753320 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8e078cc-b142-4ccb-b3e1-5c899a354113-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mtv5w\" (UID: \"d8e078cc-b142-4ccb-b3e1-5c899a354113\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753344 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4htsc\" (UID: \"2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753354 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8e078cc-b142-4ccb-b3e1-5c899a354113-service-ca-bundle\") pod \"authentication-operator-69f744f599-mtv5w\" (UID: \"d8e078cc-b142-4ccb-b3e1-5c899a354113\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753374 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-config\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753399 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5a019c2c-939e-4d88-b662-79dd417dce18-encryption-config\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753429 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhs8r\" (UniqueName: \"kubernetes.io/projected/22ccfdc9-0d00-475d-ac54-d084beaca9dd-kube-api-access-jhs8r\") pod \"downloads-7954f5f757-pmdpb\" (UID: \"22ccfdc9-0d00-475d-ac54-d084beaca9dd\") " pod="openshift-console/downloads-7954f5f757-pmdpb" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753472 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts2rz\" (UniqueName: \"kubernetes.io/projected/d62437d3-3b65-4029-9f01-f6a696e40895-kube-api-access-ts2rz\") pod \"cluster-samples-operator-665b6dd947-n7tvs\" (UID: \"d62437d3-3b65-4029-9f01-f6a696e40895\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n7tvs" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.753497 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-n757x\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.754525 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-audit-dir\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.755020 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6b18452c-12ba-4d43-b858-4608001e76d7-auth-proxy-config\") pod \"machine-approver-56656f9798-jhnjt\" (UID: \"6b18452c-12ba-4d43-b858-4608001e76d7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.755414 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.755672 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/041d3d0d-70e0-4249-a7e2-63958513ba13-config\") pod \"route-controller-manager-6576b87f9c-lqhjg\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.756081 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.756470 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.756635 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b18452c-12ba-4d43-b858-4608001e76d7-config\") pod \"machine-approver-56656f9798-jhnjt\" (UID: \"6b18452c-12ba-4d43-b858-4608001e76d7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.756886 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.757007 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sfn66"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.757021 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.757030 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.757109 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.757849 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.758244 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6b18452c-12ba-4d43-b858-4608001e76d7-machine-approver-tls\") pod \"machine-approver-56656f9798-jhnjt\" (UID: \"6b18452c-12ba-4d43-b858-4608001e76d7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.758356 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/041d3d0d-70e0-4249-a7e2-63958513ba13-client-ca\") pod \"route-controller-manager-6576b87f9c-lqhjg\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.758436 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.758572 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0d3cbb8-cbf0-4799-b40f-f97beb16a022-config\") pod \"openshift-apiserver-operator-796bbdcf4f-6qvr7\" (UID: \"f0d3cbb8-cbf0-4799-b40f-f97beb16a022\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.758936 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5a019c2c-939e-4d88-b662-79dd417dce18-audit\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.759051 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.759067 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85a7c94e-bc28-4d2e-ad77-7f776252b482-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wthrm\" (UID: \"85a7c94e-bc28-4d2e-ad77-7f776252b482\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.759096 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4-config\") pod \"machine-api-operator-5694c8668f-4htsc\" (UID: \"2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.759134 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4-images\") pod \"machine-api-operator-5694c8668f-4htsc\" (UID: \"2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.759156 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tnh7\" (UniqueName: \"kubernetes.io/projected/15377bb0-a5a9-4a17-bd32-32795886deac-kube-api-access-9tnh7\") pod \"dns-operator-744455d44c-tk4v5\" (UID: \"15377bb0-a5a9-4a17-bd32-32795886deac\") " pod="openshift-dns-operator/dns-operator-744455d44c-tk4v5" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.759174 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-audit-policies\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.759192 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46zml\" (UniqueName: \"kubernetes.io/projected/ccc071ec-9724-41f2-a654-625aae8bd840-kube-api-access-46zml\") pod \"controller-manager-879f6c89f-n757x\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.759194 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d8e078cc-b142-4ccb-b3e1-5c899a354113-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-mtv5w\" (UID: \"d8e078cc-b142-4ccb-b3e1-5c899a354113\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.760015 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.760030 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4-images\") pod \"machine-api-operator-5694c8668f-4htsc\" (UID: \"2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.760071 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-n58xw"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.760122 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8e078cc-b142-4ccb-b3e1-5c899a354113-config\") pod \"authentication-operator-69f744f599-mtv5w\" (UID: \"d8e078cc-b142-4ccb-b3e1-5c899a354113\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.760172 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4-config\") pod \"machine-api-operator-5694c8668f-4htsc\" (UID: \"2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.776118 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-serving-cert\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.776204 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-4htsc\" (UID: \"2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.776440 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-encryption-config\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.776571 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0d3cbb8-cbf0-4799-b40f-f97beb16a022-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-6qvr7\" (UID: \"f0d3cbb8-cbf0-4799-b40f-f97beb16a022\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.776895 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-etcd-client\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.777132 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8e078cc-b142-4ccb-b3e1-5c899a354113-serving-cert\") pod \"authentication-operator-69f744f599-mtv5w\" (UID: \"d8e078cc-b142-4ccb-b3e1-5c899a354113\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.760199 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/041d3d0d-70e0-4249-a7e2-63958513ba13-serving-cert\") pod \"route-controller-manager-6576b87f9c-lqhjg\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.777988 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n58xw" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.778613 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.780214 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.780451 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.783694 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.784518 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.785064 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.785389 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.786529 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.787511 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.788097 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/041d3d0d-70e0-4249-a7e2-63958513ba13-serving-cert\") pod \"route-controller-manager-6576b87f9c-lqhjg\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.788532 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.788649 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6jbrv"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.789710 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6jbrv" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.790149 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.791188 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.791482 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdd9f"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.792667 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.793078 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gkb4d"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.795215 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-gkb4d" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.796407 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.798451 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.798615 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-l9nh4"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.799387 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-l9nh4" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.800006 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zq7lm"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.804168 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.804524 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.807525 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n7tvs"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.807579 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zq7lm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.809199 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mh86q"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.811374 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-tk4v5"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.813598 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-pmdpb"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.815128 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4htsc"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.816427 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vwnln"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.817358 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6gvqj"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.818563 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.819102 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.820058 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-pcf5t"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.821065 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wpxzm"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.822039 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.823150 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.824617 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n757x"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.824807 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.825876 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.826870 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mtv5w"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.827862 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-n58xw"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.828864 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.829829 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.830801 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.833615 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.834752 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.836198 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.837277 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.837381 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.838560 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.839700 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.840828 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.841847 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cmk67"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.842914 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6jbrv"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.844102 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdd9f"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.844231 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.845814 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6gvqj"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.846872 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.848000 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zq7lm"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.849113 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-l9nh4"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.850158 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.851286 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-jsclc"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.852178 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jsclc" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.852419 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gkb4d"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.853458 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-74x48"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.855365 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jsclc"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.855479 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-74x48" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.856239 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-74x48"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.857530 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-kk9lt"] Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.858513 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-kk9lt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.864790 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.879440 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.879501 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-webhook-cert\") pod \"packageserver-d55dfcdfc-bfc6m\" (UID: \"ec5d1cfb-ed76-42f0-a140-447dd49aadb4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.879550 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-service-ca\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.879586 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ef99886-22f3-4634-9715-0dd2ee3c53a7-trusted-ca\") pod \"console-operator-58897d9998-vwnln\" (UID: \"5ef99886-22f3-4634-9715-0dd2ee3c53a7\") " pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.879621 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlgcn\" (UniqueName: \"kubernetes.io/projected/3c42e876-8a3c-40ce-afc5-eb1599c6edd1-kube-api-access-rlgcn\") pod \"olm-operator-6b444d44fb-4j9b6\" (UID: \"3c42e876-8a3c-40ce-afc5-eb1599c6edd1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.879670 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-trusted-ca-bundle\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.879703 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5a019c2c-939e-4d88-b662-79dd417dce18-node-pullsecrets\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.879775 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsxpp\" (UniqueName: \"kubernetes.io/projected/db61ff51-5c14-43bd-9042-cf7447ab7a3d-kube-api-access-tsxpp\") pod \"marketplace-operator-79b997595-rdd9f\" (UID: \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.880731 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z495x\" (UniqueName: \"kubernetes.io/projected/0b33acfc-c22d-4d1f-a70f-410ed90474bb-kube-api-access-z495x\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.880844 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c82x\" (UniqueName: \"kubernetes.io/projected/3e56bc10-6d63-4b7b-a740-b449267b3c20-kube-api-access-4c82x\") pod \"cluster-image-registry-operator-dc59b4c8b-fjz74\" (UID: \"3e56bc10-6d63-4b7b-a740-b449267b3c20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.881409 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-serving-cert\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.881481 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-service-ca\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.881492 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85a7c94e-bc28-4d2e-ad77-7f776252b482-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wthrm\" (UID: \"85a7c94e-bc28-4d2e-ad77-7f776252b482\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.881532 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/5a019c2c-939e-4d88-b662-79dd417dce18-node-pullsecrets\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.881594 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6bb8642-f0de-4683-acb3-a50a7492b9a8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6cxpz\" (UID: \"e6bb8642-f0de-4683-acb3-a50a7492b9a8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.881671 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/15377bb0-a5a9-4a17-bd32-32795886deac-metrics-tls\") pod \"dns-operator-744455d44c-tk4v5\" (UID: \"15377bb0-a5a9-4a17-bd32-32795886deac\") " pod="openshift-dns-operator/dns-operator-744455d44c-tk4v5" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.881722 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7bb99172-24fe-4dce-8334-e01d2ff734a2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-nlsd7\" (UID: \"7bb99172-24fe-4dce-8334-e01d2ff734a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.881790 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-client-ca\") pod \"controller-manager-879f6c89f-n757x\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.881844 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b5c61773-ee25-4653-abdd-304afce58b14-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fjnvq\" (UID: \"b5c61773-ee25-4653-abdd-304afce58b14\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.881906 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0b33acfc-c22d-4d1f-a70f-410ed90474bb-audit-dir\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.881986 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a019c2c-939e-4d88-b662-79dd417dce18-serving-cert\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.882040 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a019c2c-939e-4d88-b662-79dd417dce18-config\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.882104 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/d62437d3-3b65-4029-9f01-f6a696e40895-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-n7tvs\" (UID: \"d62437d3-3b65-4029-9f01-f6a696e40895\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n7tvs" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.882184 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.882249 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc6ht\" (UniqueName: \"kubernetes.io/projected/85a7c94e-bc28-4d2e-ad77-7f776252b482-kube-api-access-tc6ht\") pod \"openshift-controller-manager-operator-756b6f6bc6-wthrm\" (UID: \"85a7c94e-bc28-4d2e-ad77-7f776252b482\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.882313 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0b33acfc-c22d-4d1f-a70f-410ed90474bb-audit-dir\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.882318 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b95cf91-024d-4571-a28f-b633385428c2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-v74zl\" (UID: \"4b95cf91-024d-4571-a28f-b633385428c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.882498 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/02d20ecf-4ea9-47e2-b339-7e75e0b8e358-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr7hv\" (UID: \"02d20ecf-4ea9-47e2-b339-7e75e0b8e358\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.882606 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5ef99886-22f3-4634-9715-0dd2ee3c53a7-trusted-ca\") pod \"console-operator-58897d9998-vwnln\" (UID: \"5ef99886-22f3-4634-9715-0dd2ee3c53a7\") " pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.882640 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a1b20a8b-1487-4658-b022-1112ff5bbe8f-profile-collector-cert\") pod \"catalog-operator-68c6474976-2txch\" (UID: \"a1b20a8b-1487-4658-b022-1112ff5bbe8f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.882724 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c7c49124-758b-4fab-8567-7bd70b30feb5-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wkddb\" (UID: \"c7c49124-758b-4fab-8567-7bd70b30feb5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.882827 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5a019c2c-939e-4d88-b662-79dd417dce18-audit\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.882900 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kzkw\" (UniqueName: \"kubernetes.io/projected/7bb99172-24fe-4dce-8334-e01d2ff734a2-kube-api-access-9kzkw\") pod \"ingress-operator-5b745b69d9-nlsd7\" (UID: \"7bb99172-24fe-4dce-8334-e01d2ff734a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.882843 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-trusted-ca-bundle\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.883044 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02d20ecf-4ea9-47e2-b339-7e75e0b8e358-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr7hv\" (UID: \"02d20ecf-4ea9-47e2-b339-7e75e0b8e358\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.883174 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b95cf91-024d-4571-a28f-b633385428c2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-v74zl\" (UID: \"4b95cf91-024d-4571-a28f-b633385428c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.883318 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ef99886-22f3-4634-9715-0dd2ee3c53a7-serving-cert\") pod \"console-operator-58897d9998-vwnln\" (UID: \"5ef99886-22f3-4634-9715-0dd2ee3c53a7\") " pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.883406 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-secret-volume\") pod \"collect-profiles-29410650-wwcg4\" (UID: \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.883489 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bcb5ff5a-66f7-4bd1-9827-1bb544146761-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-l9nh4\" (UID: \"bcb5ff5a-66f7-4bd1-9827-1bb544146761\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l9nh4" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.883566 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96x9n\" (UniqueName: \"kubernetes.io/projected/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-kube-api-access-96x9n\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.883644 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj5lw\" (UniqueName: \"kubernetes.io/projected/3b0da047-6c65-43c9-b8c9-5fef3c6a7929-kube-api-access-rj5lw\") pod \"openshift-config-operator-7777fb866f-bbcm8\" (UID: \"3b0da047-6c65-43c9-b8c9-5fef3c6a7929\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.883730 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxpp8\" (UniqueName: \"kubernetes.io/projected/b5c61773-ee25-4653-abdd-304afce58b14-kube-api-access-wxpp8\") pod \"machine-config-operator-74547568cd-fjnvq\" (UID: \"b5c61773-ee25-4653-abdd-304afce58b14\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.883863 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e-config\") pod \"service-ca-operator-777779d784-n58xw\" (UID: \"d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n58xw" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.883944 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.884029 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.884136 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccc071ec-9724-41f2-a654-625aae8bd840-serving-cert\") pod \"controller-manager-879f6c89f-n757x\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.884220 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fncgp\" (UniqueName: \"kubernetes.io/projected/c7c49124-758b-4fab-8567-7bd70b30feb5-kube-api-access-fncgp\") pod \"package-server-manager-789f6589d5-wkddb\" (UID: \"c7c49124-758b-4fab-8567-7bd70b30feb5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.884351 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rdd9f\" (UID: \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.884538 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzv9t\" (UniqueName: \"kubernetes.io/projected/5ef99886-22f3-4634-9715-0dd2ee3c53a7-kube-api-access-mzv9t\") pod \"console-operator-58897d9998-vwnln\" (UID: \"5ef99886-22f3-4634-9715-0dd2ee3c53a7\") " pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.884610 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7bb99172-24fe-4dce-8334-e01d2ff734a2-metrics-tls\") pod \"ingress-operator-5b745b69d9-nlsd7\" (UID: \"7bb99172-24fe-4dce-8334-e01d2ff734a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.884728 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5a019c2c-939e-4d88-b662-79dd417dce18-image-import-ca\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.884835 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-client-ca\") pod \"controller-manager-879f6c89f-n757x\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.884861 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rdd9f\" (UID: \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.885063 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e56bc10-6d63-4b7b-a740-b449267b3c20-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fjz74\" (UID: \"3e56bc10-6d63-4b7b-a740-b449267b3c20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.885195 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3c42e876-8a3c-40ce-afc5-eb1599c6edd1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4j9b6\" (UID: \"3c42e876-8a3c-40ce-afc5-eb1599c6edd1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.885297 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frqw9\" (UniqueName: \"kubernetes.io/projected/4517cacd-a375-4c43-a8b5-7d5bf35b172a-kube-api-access-frqw9\") pod \"migrator-59844c95c7-6jbrv\" (UID: \"4517cacd-a375-4c43-a8b5-7d5bf35b172a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6jbrv" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.885376 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52dd5b9f-cc6f-4898-a5dd-a235270c39e0-service-ca-bundle\") pod \"router-default-5444994796-664p9\" (UID: \"52dd5b9f-cc6f-4898-a5dd-a235270c39e0\") " pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.885463 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.885659 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85a7c94e-bc28-4d2e-ad77-7f776252b482-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wthrm\" (UID: \"85a7c94e-bc28-4d2e-ad77-7f776252b482\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.885792 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.885875 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.885964 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/12396794-f15b-4c87-871b-55094944deb0-proxy-tls\") pod \"machine-config-controller-84d6567774-rwvbr\" (UID: \"12396794-f15b-4c87-871b-55094944deb0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.886096 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a1b20a8b-1487-4658-b022-1112ff5bbe8f-srv-cert\") pod \"catalog-operator-68c6474976-2txch\" (UID: \"a1b20a8b-1487-4658-b022-1112ff5bbe8f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.886177 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbrq7\" (UniqueName: \"kubernetes.io/projected/a1b20a8b-1487-4658-b022-1112ff5bbe8f-kube-api-access-zbrq7\") pod \"catalog-operator-68c6474976-2txch\" (UID: \"a1b20a8b-1487-4658-b022-1112ff5bbe8f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.886290 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2m45\" (UniqueName: \"kubernetes.io/projected/5a019c2c-939e-4d88-b662-79dd417dce18-kube-api-access-w2m45\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.886393 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlw5j\" (UniqueName: \"kubernetes.io/projected/bcb5ff5a-66f7-4bd1-9827-1bb544146761-kube-api-access-zlw5j\") pod \"multus-admission-controller-857f4d67dd-l9nh4\" (UID: \"bcb5ff5a-66f7-4bd1-9827-1bb544146761\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l9nh4" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.886519 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-serving-cert\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.885436 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.886627 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e56bc10-6d63-4b7b-a740-b449267b3c20-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fjz74\" (UID: \"3e56bc10-6d63-4b7b-a740-b449267b3c20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.887137 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.887145 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/5a019c2c-939e-4d88-b662-79dd417dce18-audit\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.887540 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a019c2c-939e-4d88-b662-79dd417dce18-serving-cert\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.883741 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a019c2c-939e-4d88-b662-79dd417dce18-config\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.883634 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.887919 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.888421 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/5a019c2c-939e-4d88-b662-79dd417dce18-image-import-ca\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.888985 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.889268 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmch8\" (UniqueName: \"kubernetes.io/projected/4b95cf91-024d-4571-a28f-b633385428c2-kube-api-access-vmch8\") pod \"kube-storage-version-migrator-operator-b67b599dd-v74zl\" (UID: \"4b95cf91-024d-4571-a28f-b633385428c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.889393 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5a019c2c-939e-4d88-b662-79dd417dce18-etcd-serving-ca\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.889438 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a019c2c-939e-4d88-b662-79dd417dce18-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.889522 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b5c61773-ee25-4653-abdd-304afce58b14-images\") pod \"machine-config-operator-74547568cd-fjnvq\" (UID: \"b5c61773-ee25-4653-abdd-304afce58b14\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.890236 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.890298 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5a019c2c-939e-4d88-b662-79dd417dce18-etcd-serving-ca\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.890709 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-oauth-config\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.890833 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e-serving-cert\") pod \"service-ca-operator-777779d784-n58xw\" (UID: \"d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n58xw" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.890914 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5a019c2c-939e-4d88-b662-79dd417dce18-audit-dir\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.890998 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccc071ec-9724-41f2-a654-625aae8bd840-serving-cert\") pod \"controller-manager-879f6c89f-n757x\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.891100 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b5c61773-ee25-4653-abdd-304afce58b14-proxy-tls\") pod \"machine-config-operator-74547568cd-fjnvq\" (UID: \"b5c61773-ee25-4653-abdd-304afce58b14\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.891247 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5a019c2c-939e-4d88-b662-79dd417dce18-audit-dir\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.891450 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5a019c2c-939e-4d88-b662-79dd417dce18-trusted-ca-bundle\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.891643 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-config\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.891692 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5a019c2c-939e-4d88-b662-79dd417dce18-encryption-config\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.891866 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3e56bc10-6d63-4b7b-a740-b449267b3c20-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-fjz74\" (UID: \"3e56bc10-6d63-4b7b-a740-b449267b3c20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.892457 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.892582 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6bb8642-f0de-4683-acb3-a50a7492b9a8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6cxpz\" (UID: \"e6bb8642-f0de-4683-acb3-a50a7492b9a8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.892682 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhs8r\" (UniqueName: \"kubernetes.io/projected/22ccfdc9-0d00-475d-ac54-d084beaca9dd-kube-api-access-jhs8r\") pod \"downloads-7954f5f757-pmdpb\" (UID: \"22ccfdc9-0d00-475d-ac54-d084beaca9dd\") " pod="openshift-console/downloads-7954f5f757-pmdpb" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.892734 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/d62437d3-3b65-4029-9f01-f6a696e40895-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-n7tvs\" (UID: \"d62437d3-3b65-4029-9f01-f6a696e40895\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n7tvs" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.892874 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts2rz\" (UniqueName: \"kubernetes.io/projected/d62437d3-3b65-4029-9f01-f6a696e40895-kube-api-access-ts2rz\") pod \"cluster-samples-operator-665b6dd947-n7tvs\" (UID: \"d62437d3-3b65-4029-9f01-f6a696e40895\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n7tvs" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.892886 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ef99886-22f3-4634-9715-0dd2ee3c53a7-serving-cert\") pod \"console-operator-58897d9998-vwnln\" (UID: \"5ef99886-22f3-4634-9715-0dd2ee3c53a7\") " pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.893035 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-n757x\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.893361 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-config\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.893415 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85a7c94e-bc28-4d2e-ad77-7f776252b482-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wthrm\" (UID: \"85a7c94e-bc28-4d2e-ad77-7f776252b482\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.893483 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46zml\" (UniqueName: \"kubernetes.io/projected/ccc071ec-9724-41f2-a654-625aae8bd840-kube-api-access-46zml\") pod \"controller-manager-879f6c89f-n757x\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.893574 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/52dd5b9f-cc6f-4898-a5dd-a235270c39e0-stats-auth\") pod \"router-default-5444994796-664p9\" (UID: \"52dd5b9f-cc6f-4898-a5dd-a235270c39e0\") " pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.893731 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tnh7\" (UniqueName: \"kubernetes.io/projected/15377bb0-a5a9-4a17-bd32-32795886deac-kube-api-access-9tnh7\") pod \"dns-operator-744455d44c-tk4v5\" (UID: \"15377bb0-a5a9-4a17-bd32-32795886deac\") " pod="openshift-dns-operator/dns-operator-744455d44c-tk4v5" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.893846 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-audit-policies\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.893947 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77cbq\" (UniqueName: \"kubernetes.io/projected/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-kube-api-access-77cbq\") pod \"collect-profiles-29410650-wwcg4\" (UID: \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.894188 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-oauth-config\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.894080 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/12396794-f15b-4c87-871b-55094944deb0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rwvbr\" (UID: \"12396794-f15b-4c87-871b-55094944deb0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.894365 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85a7c94e-bc28-4d2e-ad77-7f776252b482-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wthrm\" (UID: \"85a7c94e-bc28-4d2e-ad77-7f776252b482\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.894369 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwt4x\" (UniqueName: \"kubernetes.io/projected/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-kube-api-access-jwt4x\") pod \"packageserver-d55dfcdfc-bfc6m\" (UID: \"ec5d1cfb-ed76-42f0-a140-447dd49aadb4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.894606 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.894689 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-config\") pod \"controller-manager-879f6c89f-n757x\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.894854 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02d20ecf-4ea9-47e2-b339-7e75e0b8e358-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr7hv\" (UID: \"02d20ecf-4ea9-47e2-b339-7e75e0b8e358\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.894959 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ef99886-22f3-4634-9715-0dd2ee3c53a7-config\") pod \"console-operator-58897d9998-vwnln\" (UID: \"5ef99886-22f3-4634-9715-0dd2ee3c53a7\") " pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.895031 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6bb8642-f0de-4683-acb3-a50a7492b9a8-config\") pod \"kube-apiserver-operator-766d6c64bb-6cxpz\" (UID: \"e6bb8642-f0de-4683-acb3-a50a7492b9a8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.895156 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-config-volume\") pod \"collect-profiles-29410650-wwcg4\" (UID: \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.895241 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5a019c2c-939e-4d88-b662-79dd417dce18-etcd-client\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.895330 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.895405 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.895502 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3b0da047-6c65-43c9-b8c9-5fef3c6a7929-available-featuregates\") pod \"openshift-config-operator-7777fb866f-bbcm8\" (UID: \"3b0da047-6c65-43c9-b8c9-5fef3c6a7929\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.895578 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7bb99172-24fe-4dce-8334-e01d2ff734a2-trusted-ca\") pod \"ingress-operator-5b745b69d9-nlsd7\" (UID: \"7bb99172-24fe-4dce-8334-e01d2ff734a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.894567 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-audit-policies\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.895716 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.895803 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-apiservice-cert\") pod \"packageserver-d55dfcdfc-bfc6m\" (UID: \"ec5d1cfb-ed76-42f0-a140-447dd49aadb4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.895876 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/52dd5b9f-cc6f-4898-a5dd-a235270c39e0-default-certificate\") pod \"router-default-5444994796-664p9\" (UID: \"52dd5b9f-cc6f-4898-a5dd-a235270c39e0\") " pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.895968 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-tmpfs\") pod \"packageserver-d55dfcdfc-bfc6m\" (UID: \"ec5d1cfb-ed76-42f0-a140-447dd49aadb4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.896047 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/52dd5b9f-cc6f-4898-a5dd-a235270c39e0-metrics-certs\") pod \"router-default-5444994796-664p9\" (UID: \"52dd5b9f-cc6f-4898-a5dd-a235270c39e0\") " pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.896125 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3e56bc10-6d63-4b7b-a740-b449267b3c20-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fjz74\" (UID: \"3e56bc10-6d63-4b7b-a740-b449267b3c20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.896195 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b0da047-6c65-43c9-b8c9-5fef3c6a7929-serving-cert\") pod \"openshift-config-operator-7777fb866f-bbcm8\" (UID: \"3b0da047-6c65-43c9-b8c9-5fef3c6a7929\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.896266 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3c42e876-8a3c-40ce-afc5-eb1599c6edd1-srv-cert\") pod \"olm-operator-6b444d44fb-4j9b6\" (UID: \"3c42e876-8a3c-40ce-afc5-eb1599c6edd1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.896348 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-oauth-serving-cert\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.896771 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbmvd\" (UniqueName: \"kubernetes.io/projected/12396794-f15b-4c87-871b-55094944deb0-kube-api-access-pbmvd\") pod \"machine-config-controller-84d6567774-rwvbr\" (UID: \"12396794-f15b-4c87-871b-55094944deb0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.896873 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xrls\" (UniqueName: \"kubernetes.io/projected/d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e-kube-api-access-7xrls\") pod \"service-ca-operator-777779d784-n58xw\" (UID: \"d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n58xw" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.896947 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmggs\" (UniqueName: \"kubernetes.io/projected/52dd5b9f-cc6f-4898-a5dd-a235270c39e0-kube-api-access-mmggs\") pod \"router-default-5444994796-664p9\" (UID: \"52dd5b9f-cc6f-4898-a5dd-a235270c39e0\") " pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.896014 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-n757x\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.897283 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ef99886-22f3-4634-9715-0dd2ee3c53a7-config\") pod \"console-operator-58897d9998-vwnln\" (UID: \"5ef99886-22f3-4634-9715-0dd2ee3c53a7\") " pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.897890 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/3b0da047-6c65-43c9-b8c9-5fef3c6a7929-available-featuregates\") pod \"openshift-config-operator-7777fb866f-bbcm8\" (UID: \"3b0da047-6c65-43c9-b8c9-5fef3c6a7929\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.898130 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.896304 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/15377bb0-a5a9-4a17-bd32-32795886deac-metrics-tls\") pod \"dns-operator-744455d44c-tk4v5\" (UID: \"15377bb0-a5a9-4a17-bd32-32795886deac\") " pod="openshift-dns-operator/dns-operator-744455d44c-tk4v5" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.898910 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-config\") pod \"controller-manager-879f6c89f-n757x\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.899681 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-oauth-serving-cert\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.899988 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.900738 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3e56bc10-6d63-4b7b-a740-b449267b3c20-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-fjz74\" (UID: \"3e56bc10-6d63-4b7b-a740-b449267b3c20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.900919 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.901002 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5a019c2c-939e-4d88-b662-79dd417dce18-encryption-config\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.901044 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5a019c2c-939e-4d88-b662-79dd417dce18-etcd-client\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.902164 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b0da047-6c65-43c9-b8c9-5fef3c6a7929-serving-cert\") pod \"openshift-config-operator-7777fb866f-bbcm8\" (UID: \"3b0da047-6c65-43c9-b8c9-5fef3c6a7929\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.902616 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.902880 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.907152 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.925620 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.945772 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.965404 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.985167 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.998380 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rdd9f\" (UID: \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.998464 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3c42e876-8a3c-40ce-afc5-eb1599c6edd1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4j9b6\" (UID: \"3c42e876-8a3c-40ce-afc5-eb1599c6edd1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.998521 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frqw9\" (UniqueName: \"kubernetes.io/projected/4517cacd-a375-4c43-a8b5-7d5bf35b172a-kube-api-access-frqw9\") pod \"migrator-59844c95c7-6jbrv\" (UID: \"4517cacd-a375-4c43-a8b5-7d5bf35b172a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6jbrv" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.998554 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52dd5b9f-cc6f-4898-a5dd-a235270c39e0-service-ca-bundle\") pod \"router-default-5444994796-664p9\" (UID: \"52dd5b9f-cc6f-4898-a5dd-a235270c39e0\") " pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.998610 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/12396794-f15b-4c87-871b-55094944deb0-proxy-tls\") pod \"machine-config-controller-84d6567774-rwvbr\" (UID: \"12396794-f15b-4c87-871b-55094944deb0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.998636 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a1b20a8b-1487-4658-b022-1112ff5bbe8f-srv-cert\") pod \"catalog-operator-68c6474976-2txch\" (UID: \"a1b20a8b-1487-4658-b022-1112ff5bbe8f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.998660 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbrq7\" (UniqueName: \"kubernetes.io/projected/a1b20a8b-1487-4658-b022-1112ff5bbe8f-kube-api-access-zbrq7\") pod \"catalog-operator-68c6474976-2txch\" (UID: \"a1b20a8b-1487-4658-b022-1112ff5bbe8f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.998697 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmch8\" (UniqueName: \"kubernetes.io/projected/4b95cf91-024d-4571-a28f-b633385428c2-kube-api-access-vmch8\") pod \"kube-storage-version-migrator-operator-b67b599dd-v74zl\" (UID: \"4b95cf91-024d-4571-a28f-b633385428c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.998720 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlw5j\" (UniqueName: \"kubernetes.io/projected/bcb5ff5a-66f7-4bd1-9827-1bb544146761-kube-api-access-zlw5j\") pod \"multus-admission-controller-857f4d67dd-l9nh4\" (UID: \"bcb5ff5a-66f7-4bd1-9827-1bb544146761\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l9nh4" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.998808 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b5c61773-ee25-4653-abdd-304afce58b14-images\") pod \"machine-config-operator-74547568cd-fjnvq\" (UID: \"b5c61773-ee25-4653-abdd-304afce58b14\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.998847 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e-serving-cert\") pod \"service-ca-operator-777779d784-n58xw\" (UID: \"d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n58xw" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.998870 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b5c61773-ee25-4653-abdd-304afce58b14-proxy-tls\") pod \"machine-config-operator-74547568cd-fjnvq\" (UID: \"b5c61773-ee25-4653-abdd-304afce58b14\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.998922 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6bb8642-f0de-4683-acb3-a50a7492b9a8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6cxpz\" (UID: \"e6bb8642-f0de-4683-acb3-a50a7492b9a8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.998959 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/52dd5b9f-cc6f-4898-a5dd-a235270c39e0-stats-auth\") pod \"router-default-5444994796-664p9\" (UID: \"52dd5b9f-cc6f-4898-a5dd-a235270c39e0\") " pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.998998 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/12396794-f15b-4c87-871b-55094944deb0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rwvbr\" (UID: \"12396794-f15b-4c87-871b-55094944deb0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999042 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwt4x\" (UniqueName: \"kubernetes.io/projected/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-kube-api-access-jwt4x\") pod \"packageserver-d55dfcdfc-bfc6m\" (UID: \"ec5d1cfb-ed76-42f0-a140-447dd49aadb4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999074 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77cbq\" (UniqueName: \"kubernetes.io/projected/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-kube-api-access-77cbq\") pod \"collect-profiles-29410650-wwcg4\" (UID: \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999102 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02d20ecf-4ea9-47e2-b339-7e75e0b8e358-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr7hv\" (UID: \"02d20ecf-4ea9-47e2-b339-7e75e0b8e358\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999138 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6bb8642-f0de-4683-acb3-a50a7492b9a8-config\") pod \"kube-apiserver-operator-766d6c64bb-6cxpz\" (UID: \"e6bb8642-f0de-4683-acb3-a50a7492b9a8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999161 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-config-volume\") pod \"collect-profiles-29410650-wwcg4\" (UID: \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999221 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7bb99172-24fe-4dce-8334-e01d2ff734a2-trusted-ca\") pod \"ingress-operator-5b745b69d9-nlsd7\" (UID: \"7bb99172-24fe-4dce-8334-e01d2ff734a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999250 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-apiservice-cert\") pod \"packageserver-d55dfcdfc-bfc6m\" (UID: \"ec5d1cfb-ed76-42f0-a140-447dd49aadb4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999274 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/52dd5b9f-cc6f-4898-a5dd-a235270c39e0-default-certificate\") pod \"router-default-5444994796-664p9\" (UID: \"52dd5b9f-cc6f-4898-a5dd-a235270c39e0\") " pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999307 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/52dd5b9f-cc6f-4898-a5dd-a235270c39e0-metrics-certs\") pod \"router-default-5444994796-664p9\" (UID: \"52dd5b9f-cc6f-4898-a5dd-a235270c39e0\") " pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999336 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-tmpfs\") pod \"packageserver-d55dfcdfc-bfc6m\" (UID: \"ec5d1cfb-ed76-42f0-a140-447dd49aadb4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999372 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3c42e876-8a3c-40ce-afc5-eb1599c6edd1-srv-cert\") pod \"olm-operator-6b444d44fb-4j9b6\" (UID: \"3c42e876-8a3c-40ce-afc5-eb1599c6edd1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999397 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbmvd\" (UniqueName: \"kubernetes.io/projected/12396794-f15b-4c87-871b-55094944deb0-kube-api-access-pbmvd\") pod \"machine-config-controller-84d6567774-rwvbr\" (UID: \"12396794-f15b-4c87-871b-55094944deb0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999420 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xrls\" (UniqueName: \"kubernetes.io/projected/d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e-kube-api-access-7xrls\") pod \"service-ca-operator-777779d784-n58xw\" (UID: \"d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n58xw" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999443 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmggs\" (UniqueName: \"kubernetes.io/projected/52dd5b9f-cc6f-4898-a5dd-a235270c39e0-kube-api-access-mmggs\") pod \"router-default-5444994796-664p9\" (UID: \"52dd5b9f-cc6f-4898-a5dd-a235270c39e0\") " pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999464 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-webhook-cert\") pod \"packageserver-d55dfcdfc-bfc6m\" (UID: \"ec5d1cfb-ed76-42f0-a140-447dd49aadb4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999488 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlgcn\" (UniqueName: \"kubernetes.io/projected/3c42e876-8a3c-40ce-afc5-eb1599c6edd1-kube-api-access-rlgcn\") pod \"olm-operator-6b444d44fb-4j9b6\" (UID: \"3c42e876-8a3c-40ce-afc5-eb1599c6edd1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999512 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsxpp\" (UniqueName: \"kubernetes.io/projected/db61ff51-5c14-43bd-9042-cf7447ab7a3d-kube-api-access-tsxpp\") pod \"marketplace-operator-79b997595-rdd9f\" (UID: \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999579 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6bb8642-f0de-4683-acb3-a50a7492b9a8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6cxpz\" (UID: \"e6bb8642-f0de-4683-acb3-a50a7492b9a8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999626 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7bb99172-24fe-4dce-8334-e01d2ff734a2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-nlsd7\" (UID: \"7bb99172-24fe-4dce-8334-e01d2ff734a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999650 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b5c61773-ee25-4653-abdd-304afce58b14-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fjnvq\" (UID: \"b5c61773-ee25-4653-abdd-304afce58b14\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999695 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b95cf91-024d-4571-a28f-b633385428c2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-v74zl\" (UID: \"4b95cf91-024d-4571-a28f-b633385428c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999719 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/02d20ecf-4ea9-47e2-b339-7e75e0b8e358-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr7hv\" (UID: \"02d20ecf-4ea9-47e2-b339-7e75e0b8e358\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999760 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a1b20a8b-1487-4658-b022-1112ff5bbe8f-profile-collector-cert\") pod \"catalog-operator-68c6474976-2txch\" (UID: \"a1b20a8b-1487-4658-b022-1112ff5bbe8f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999787 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c7c49124-758b-4fab-8567-7bd70b30feb5-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wkddb\" (UID: \"c7c49124-758b-4fab-8567-7bd70b30feb5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999814 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kzkw\" (UniqueName: \"kubernetes.io/projected/7bb99172-24fe-4dce-8334-e01d2ff734a2-kube-api-access-9kzkw\") pod \"ingress-operator-5b745b69d9-nlsd7\" (UID: \"7bb99172-24fe-4dce-8334-e01d2ff734a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999883 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02d20ecf-4ea9-47e2-b339-7e75e0b8e358-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr7hv\" (UID: \"02d20ecf-4ea9-47e2-b339-7e75e0b8e358\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999908 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b95cf91-024d-4571-a28f-b633385428c2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-v74zl\" (UID: \"4b95cf91-024d-4571-a28f-b633385428c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl" Dec 02 01:38:39 crc kubenswrapper[4884]: I1202 01:38:39.999942 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bcb5ff5a-66f7-4bd1-9827-1bb544146761-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-l9nh4\" (UID: \"bcb5ff5a-66f7-4bd1-9827-1bb544146761\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l9nh4" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:39.999990 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-secret-volume\") pod \"collect-profiles-29410650-wwcg4\" (UID: \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.000013 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxpp8\" (UniqueName: \"kubernetes.io/projected/b5c61773-ee25-4653-abdd-304afce58b14-kube-api-access-wxpp8\") pod \"machine-config-operator-74547568cd-fjnvq\" (UID: \"b5c61773-ee25-4653-abdd-304afce58b14\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.000048 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fncgp\" (UniqueName: \"kubernetes.io/projected/c7c49124-758b-4fab-8567-7bd70b30feb5-kube-api-access-fncgp\") pod \"package-server-manager-789f6589d5-wkddb\" (UID: \"c7c49124-758b-4fab-8567-7bd70b30feb5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.000071 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e-config\") pod \"service-ca-operator-777779d784-n58xw\" (UID: \"d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n58xw" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.000095 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rdd9f\" (UID: \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.000133 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7bb99172-24fe-4dce-8334-e01d2ff734a2-metrics-tls\") pod \"ingress-operator-5b745b69d9-nlsd7\" (UID: \"7bb99172-24fe-4dce-8334-e01d2ff734a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.000697 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/12396794-f15b-4c87-871b-55094944deb0-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rwvbr\" (UID: \"12396794-f15b-4c87-871b-55094944deb0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.001091 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-tmpfs\") pod \"packageserver-d55dfcdfc-bfc6m\" (UID: \"ec5d1cfb-ed76-42f0-a140-447dd49aadb4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.002100 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b5c61773-ee25-4653-abdd-304afce58b14-auth-proxy-config\") pod \"machine-config-operator-74547568cd-fjnvq\" (UID: \"b5c61773-ee25-4653-abdd-304afce58b14\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.005203 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.024691 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.044773 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.064938 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.085191 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.104959 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.115003 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/52dd5b9f-cc6f-4898-a5dd-a235270c39e0-default-certificate\") pod \"router-default-5444994796-664p9\" (UID: \"52dd5b9f-cc6f-4898-a5dd-a235270c39e0\") " pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.124919 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.135659 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/52dd5b9f-cc6f-4898-a5dd-a235270c39e0-stats-auth\") pod \"router-default-5444994796-664p9\" (UID: \"52dd5b9f-cc6f-4898-a5dd-a235270c39e0\") " pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.145164 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.164514 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.169383 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/52dd5b9f-cc6f-4898-a5dd-a235270c39e0-service-ca-bundle\") pod \"router-default-5444994796-664p9\" (UID: \"52dd5b9f-cc6f-4898-a5dd-a235270c39e0\") " pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.185969 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.196369 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/52dd5b9f-cc6f-4898-a5dd-a235270c39e0-metrics-certs\") pod \"router-default-5444994796-664p9\" (UID: \"52dd5b9f-cc6f-4898-a5dd-a235270c39e0\") " pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.204676 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.234361 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.241741 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7bb99172-24fe-4dce-8334-e01d2ff734a2-trusted-ca\") pod \"ingress-operator-5b745b69d9-nlsd7\" (UID: \"7bb99172-24fe-4dce-8334-e01d2ff734a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.245156 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.265481 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.286835 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.305837 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.325282 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.336194 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7bb99172-24fe-4dce-8334-e01d2ff734a2-metrics-tls\") pod \"ingress-operator-5b745b69d9-nlsd7\" (UID: \"7bb99172-24fe-4dce-8334-e01d2ff734a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.344306 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.365008 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.384258 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.404694 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.414240 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/3c42e876-8a3c-40ce-afc5-eb1599c6edd1-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4j9b6\" (UID: \"3c42e876-8a3c-40ce-afc5-eb1599c6edd1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.417139 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-secret-volume\") pod \"collect-profiles-29410650-wwcg4\" (UID: \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.419018 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a1b20a8b-1487-4658-b022-1112ff5bbe8f-profile-collector-cert\") pod \"catalog-operator-68c6474976-2txch\" (UID: \"a1b20a8b-1487-4658-b022-1112ff5bbe8f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.424563 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.444966 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.454627 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a1b20a8b-1487-4658-b022-1112ff5bbe8f-srv-cert\") pod \"catalog-operator-68c6474976-2txch\" (UID: \"a1b20a8b-1487-4658-b022-1112ff5bbe8f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.466216 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.512165 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rht9g\" (UniqueName: \"kubernetes.io/projected/d8e078cc-b142-4ccb-b3e1-5c899a354113-kube-api-access-rht9g\") pod \"authentication-operator-69f744f599-mtv5w\" (UID: \"d8e078cc-b142-4ccb-b3e1-5c899a354113\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.531545 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcp99\" (UniqueName: \"kubernetes.io/projected/6b18452c-12ba-4d43-b858-4608001e76d7-kube-api-access-hcp99\") pod \"machine-approver-56656f9798-jhnjt\" (UID: \"6b18452c-12ba-4d43-b858-4608001e76d7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.546621 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7v5b\" (UniqueName: \"kubernetes.io/projected/041d3d0d-70e0-4249-a7e2-63958513ba13-kube-api-access-q7v5b\") pod \"route-controller-manager-6576b87f9c-lqhjg\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.584951 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.595672 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnm4m\" (UniqueName: \"kubernetes.io/projected/f0d3cbb8-cbf0-4799-b40f-f97beb16a022-kube-api-access-lnm4m\") pod \"openshift-apiserver-operator-796bbdcf4f-6qvr7\" (UID: \"f0d3cbb8-cbf0-4799-b40f-f97beb16a022\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.608656 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.616672 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e6bb8642-f0de-4683-acb3-a50a7492b9a8-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6cxpz\" (UID: \"e6bb8642-f0de-4683-acb3-a50a7492b9a8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.625913 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.640780 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.644722 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.648593 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.653228 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e6bb8642-f0de-4683-acb3-a50a7492b9a8-config\") pod \"kube-apiserver-operator-766d6c64bb-6cxpz\" (UID: \"e6bb8642-f0de-4683-acb3-a50a7492b9a8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.687094 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.691553 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x9ws\" (UniqueName: \"kubernetes.io/projected/313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed-kube-api-access-4x9ws\") pod \"apiserver-7bbb656c7d-cswlp\" (UID: \"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.696791 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c7c49124-758b-4fab-8567-7bd70b30feb5-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wkddb\" (UID: \"c7c49124-758b-4fab-8567-7bd70b30feb5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.725191 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.734358 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smp4d\" (UniqueName: \"kubernetes.io/projected/2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4-kube-api-access-smp4d\") pod \"machine-api-operator-5694c8668f-4htsc\" (UID: \"2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.745445 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.756246 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/02d20ecf-4ea9-47e2-b339-7e75e0b8e358-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr7hv\" (UID: \"02d20ecf-4ea9-47e2-b339-7e75e0b8e358\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.763572 4884 request.go:700] Waited for 1.004245608s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-scheduler-operator/secrets?fieldSelector=metadata.name%3Dopenshift-kube-scheduler-operator-dockercfg-qt55r&limit=500&resourceVersion=0 Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.766365 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.791132 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.793383 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02d20ecf-4ea9-47e2-b339-7e75e0b8e358-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr7hv\" (UID: \"02d20ecf-4ea9-47e2-b339-7e75e0b8e358\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.805346 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.819616 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.821486 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e-serving-cert\") pod \"service-ca-operator-777779d784-n58xw\" (UID: \"d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n58xw" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.826034 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.828187 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.848007 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.866522 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.871286 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.873210 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e-config\") pod \"service-ca-operator-777779d784-n58xw\" (UID: \"d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n58xw" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.884992 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.886088 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.905807 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.915188 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/3c42e876-8a3c-40ce-afc5-eb1599c6edd1-srv-cert\") pod \"olm-operator-6b444d44fb-4j9b6\" (UID: \"3c42e876-8a3c-40ce-afc5-eb1599c6edd1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.926311 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.946321 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.966594 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.975970 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b95cf91-024d-4571-a28f-b633385428c2-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-v74zl\" (UID: \"4b95cf91-024d-4571-a28f-b633385428c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.986657 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 02 01:38:40 crc kubenswrapper[4884]: I1202 01:38:40.992569 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b95cf91-024d-4571-a28f-b633385428c2-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-v74zl\" (UID: \"4b95cf91-024d-4571-a28f-b633385428c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl" Dec 02 01:38:40 crc kubenswrapper[4884]: E1202 01:38:40.999286 4884 secret.go:188] Couldn't get secret openshift-machine-config-operator/mcc-proxy-tls: failed to sync secret cache: timed out waiting for the condition Dec 02 01:38:40 crc kubenswrapper[4884]: E1202 01:38:40.999338 4884 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Dec 02 01:38:40 crc kubenswrapper[4884]: E1202 01:38:40.999351 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/12396794-f15b-4c87-871b-55094944deb0-proxy-tls podName:12396794-f15b-4c87-871b-55094944deb0 nodeName:}" failed. No retries permitted until 2025-12-02 01:38:41.499332755 +0000 UTC m=+138.175169639 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/12396794-f15b-4c87-871b-55094944deb0-proxy-tls") pod "machine-config-controller-84d6567774-rwvbr" (UID: "12396794-f15b-4c87-871b-55094944deb0") : failed to sync secret cache: timed out waiting for the condition Dec 02 01:38:40 crc kubenswrapper[4884]: E1202 01:38:40.999429 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-trusted-ca podName:db61ff51-5c14-43bd-9042-cf7447ab7a3d nodeName:}" failed. No retries permitted until 2025-12-02 01:38:41.499411337 +0000 UTC m=+138.175248221 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-trusted-ca") pod "marketplace-operator-79b997595-rdd9f" (UID: "db61ff51-5c14-43bd-9042-cf7447ab7a3d") : failed to sync configmap cache: timed out waiting for the condition Dec 02 01:38:41 crc kubenswrapper[4884]: E1202 01:38:41.000618 4884 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Dec 02 01:38:41 crc kubenswrapper[4884]: E1202 01:38:41.000669 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-config-volume podName:83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5 nodeName:}" failed. No retries permitted until 2025-12-02 01:38:41.500643707 +0000 UTC m=+138.176480591 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-config-volume") pod "collect-profiles-29410650-wwcg4" (UID: "83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5") : failed to sync configmap cache: timed out waiting for the condition Dec 02 01:38:41 crc kubenswrapper[4884]: E1202 01:38:41.000693 4884 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Dec 02 01:38:41 crc kubenswrapper[4884]: E1202 01:38:41.000717 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-apiservice-cert podName:ec5d1cfb-ed76-42f0-a140-447dd49aadb4 nodeName:}" failed. No retries permitted until 2025-12-02 01:38:41.500711599 +0000 UTC m=+138.176548483 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-apiservice-cert") pod "packageserver-d55dfcdfc-bfc6m" (UID: "ec5d1cfb-ed76-42f0-a140-447dd49aadb4") : failed to sync secret cache: timed out waiting for the condition Dec 02 01:38:41 crc kubenswrapper[4884]: E1202 01:38:41.001137 4884 secret.go:188] Couldn't get secret openshift-machine-config-operator/mco-proxy-tls: failed to sync secret cache: timed out waiting for the condition Dec 02 01:38:41 crc kubenswrapper[4884]: E1202 01:38:41.001175 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5c61773-ee25-4653-abdd-304afce58b14-proxy-tls podName:b5c61773-ee25-4653-abdd-304afce58b14 nodeName:}" failed. No retries permitted until 2025-12-02 01:38:41.50116708 +0000 UTC m=+138.177003974 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/b5c61773-ee25-4653-abdd-304afce58b14-proxy-tls") pod "machine-config-operator-74547568cd-fjnvq" (UID: "b5c61773-ee25-4653-abdd-304afce58b14") : failed to sync secret cache: timed out waiting for the condition Dec 02 01:38:41 crc kubenswrapper[4884]: E1202 01:38:41.001218 4884 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/machine-config-operator-images: failed to sync configmap cache: timed out waiting for the condition Dec 02 01:38:41 crc kubenswrapper[4884]: E1202 01:38:41.001240 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b5c61773-ee25-4653-abdd-304afce58b14-images podName:b5c61773-ee25-4653-abdd-304afce58b14 nodeName:}" failed. No retries permitted until 2025-12-02 01:38:41.501234552 +0000 UTC m=+138.177071436 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/b5c61773-ee25-4653-abdd-304afce58b14-images") pod "machine-config-operator-74547568cd-fjnvq" (UID: "b5c61773-ee25-4653-abdd-304afce58b14") : failed to sync configmap cache: timed out waiting for the condition Dec 02 01:38:41 crc kubenswrapper[4884]: E1202 01:38:41.001820 4884 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Dec 02 01:38:41 crc kubenswrapper[4884]: E1202 01:38:41.001881 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bcb5ff5a-66f7-4bd1-9827-1bb544146761-webhook-certs podName:bcb5ff5a-66f7-4bd1-9827-1bb544146761 nodeName:}" failed. No retries permitted until 2025-12-02 01:38:41.501852447 +0000 UTC m=+138.177689331 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/bcb5ff5a-66f7-4bd1-9827-1bb544146761-webhook-certs") pod "multus-admission-controller-857f4d67dd-l9nh4" (UID: "bcb5ff5a-66f7-4bd1-9827-1bb544146761") : failed to sync secret cache: timed out waiting for the condition Dec 02 01:38:41 crc kubenswrapper[4884]: E1202 01:38:41.001900 4884 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Dec 02 01:38:41 crc kubenswrapper[4884]: E1202 01:38:41.001923 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-webhook-cert podName:ec5d1cfb-ed76-42f0-a140-447dd49aadb4 nodeName:}" failed. No retries permitted until 2025-12-02 01:38:41.501917198 +0000 UTC m=+138.177754082 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-webhook-cert") pod "packageserver-d55dfcdfc-bfc6m" (UID: "ec5d1cfb-ed76-42f0-a140-447dd49aadb4") : failed to sync secret cache: timed out waiting for the condition Dec 02 01:38:41 crc kubenswrapper[4884]: E1202 01:38:41.002349 4884 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Dec 02 01:38:41 crc kubenswrapper[4884]: E1202 01:38:41.002406 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-operator-metrics podName:db61ff51-5c14-43bd-9042-cf7447ab7a3d nodeName:}" failed. No retries permitted until 2025-12-02 01:38:41.5023805 +0000 UTC m=+138.178217384 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-operator-metrics") pod "marketplace-operator-79b997595-rdd9f" (UID: "db61ff51-5c14-43bd-9042-cf7447ab7a3d") : failed to sync secret cache: timed out waiting for the condition Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.005481 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.015147 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-mtv5w"] Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.024140 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.045082 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.051520 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg"] Dec 02 01:38:41 crc kubenswrapper[4884]: W1202 01:38:41.057308 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod041d3d0d_70e0_4249_a7e2_63958513ba13.slice/crio-2f34b84d127592f58fe99f02094b29e5bcdf83bf6f1147b79a9dfba1b4b49c2d WatchSource:0}: Error finding container 2f34b84d127592f58fe99f02094b29e5bcdf83bf6f1147b79a9dfba1b4b49c2d: Status 404 returned error can't find the container with id 2f34b84d127592f58fe99f02094b29e5bcdf83bf6f1147b79a9dfba1b4b49c2d Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.064692 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.084187 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.105149 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.117433 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7"] Dec 02 01:38:41 crc kubenswrapper[4884]: W1202 01:38:41.123597 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0d3cbb8_cbf0_4799_b40f_f97beb16a022.slice/crio-ba84b734a660bbe8df0919f81d0f7016dc19c157186d206baf40471ef17d0605 WatchSource:0}: Error finding container ba84b734a660bbe8df0919f81d0f7016dc19c157186d206baf40471ef17d0605: Status 404 returned error can't find the container with id ba84b734a660bbe8df0919f81d0f7016dc19c157186d206baf40471ef17d0605 Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.124300 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.144768 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.164621 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.185258 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.205573 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.224949 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.245448 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.275032 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.306325 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.306451 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.325369 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.339385 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-4htsc"] Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.346090 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.349885 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp"] Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.364893 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.385473 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.405098 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.424738 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.445231 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.466042 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.506324 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.525820 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.531159 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rdd9f\" (UID: \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.531298 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/12396794-f15b-4c87-871b-55094944deb0-proxy-tls\") pod \"machine-config-controller-84d6567774-rwvbr\" (UID: \"12396794-f15b-4c87-871b-55094944deb0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.531466 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b5c61773-ee25-4653-abdd-304afce58b14-images\") pod \"machine-config-operator-74547568cd-fjnvq\" (UID: \"b5c61773-ee25-4653-abdd-304afce58b14\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.531529 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b5c61773-ee25-4653-abdd-304afce58b14-proxy-tls\") pod \"machine-config-operator-74547568cd-fjnvq\" (UID: \"b5c61773-ee25-4653-abdd-304afce58b14\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.531706 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-config-volume\") pod \"collect-profiles-29410650-wwcg4\" (UID: \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.531827 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-apiservice-cert\") pod \"packageserver-d55dfcdfc-bfc6m\" (UID: \"ec5d1cfb-ed76-42f0-a140-447dd49aadb4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.531942 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-webhook-cert\") pod \"packageserver-d55dfcdfc-bfc6m\" (UID: \"ec5d1cfb-ed76-42f0-a140-447dd49aadb4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.532208 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bcb5ff5a-66f7-4bd1-9827-1bb544146761-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-l9nh4\" (UID: \"bcb5ff5a-66f7-4bd1-9827-1bb544146761\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l9nh4" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.532329 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rdd9f\" (UID: \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.532389 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rdd9f\" (UID: \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.533096 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b5c61773-ee25-4653-abdd-304afce58b14-images\") pod \"machine-config-operator-74547568cd-fjnvq\" (UID: \"b5c61773-ee25-4653-abdd-304afce58b14\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.533259 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-config-volume\") pod \"collect-profiles-29410650-wwcg4\" (UID: \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.537337 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-apiservice-cert\") pod \"packageserver-d55dfcdfc-bfc6m\" (UID: \"ec5d1cfb-ed76-42f0-a140-447dd49aadb4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.537716 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rdd9f\" (UID: \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.538293 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/bcb5ff5a-66f7-4bd1-9827-1bb544146761-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-l9nh4\" (UID: \"bcb5ff5a-66f7-4bd1-9827-1bb544146761\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l9nh4" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.538769 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-webhook-cert\") pod \"packageserver-d55dfcdfc-bfc6m\" (UID: \"ec5d1cfb-ed76-42f0-a140-447dd49aadb4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.539177 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b5c61773-ee25-4653-abdd-304afce58b14-proxy-tls\") pod \"machine-config-operator-74547568cd-fjnvq\" (UID: \"b5c61773-ee25-4653-abdd-304afce58b14\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.540290 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/12396794-f15b-4c87-871b-55094944deb0-proxy-tls\") pod \"machine-config-controller-84d6567774-rwvbr\" (UID: \"12396794-f15b-4c87-871b-55094944deb0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.545464 4884 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.564490 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.585508 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.604856 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.625374 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" event={"ID":"041d3d0d-70e0-4249-a7e2-63958513ba13","Type":"ContainerStarted","Data":"2f34b84d127592f58fe99f02094b29e5bcdf83bf6f1147b79a9dfba1b4b49c2d"} Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.625447 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" event={"ID":"d8e078cc-b142-4ccb-b3e1-5c899a354113","Type":"ContainerStarted","Data":"5bf6cb3cf4978247c4a8fbf2ca85da19daaaf9504fe015a3b0e0b73aff2d8f7f"} Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.625478 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" event={"ID":"6b18452c-12ba-4d43-b858-4608001e76d7","Type":"ContainerStarted","Data":"5a7a66deb8b8f7d62d1e1cf3dc1f7ec12281716c5c3fcf8362d06ffade6fd314"} Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.625503 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7" event={"ID":"f0d3cbb8-cbf0-4799-b40f-f97beb16a022","Type":"ContainerStarted","Data":"ba84b734a660bbe8df0919f81d0f7016dc19c157186d206baf40471ef17d0605"} Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.625979 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.646419 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.665611 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.685356 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.705362 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.725655 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 02 01:38:41 crc kubenswrapper[4884]: W1202 01:38:41.730428 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b7e2531_5e19_4cf0_9ff1_20e13b3db8b4.slice/crio-02d235cb9182865ba2833b750b4da501237067e994611b998f841ec8d4aea93b WatchSource:0}: Error finding container 02d235cb9182865ba2833b750b4da501237067e994611b998f841ec8d4aea93b: Status 404 returned error can't find the container with id 02d235cb9182865ba2833b750b4da501237067e994611b998f841ec8d4aea93b Dec 02 01:38:41 crc kubenswrapper[4884]: W1202 01:38:41.732364 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod313b4c60_5ac8_40a8_a2b4_82ab23f5c8ed.slice/crio-a15d77b97c9b3f13f9ce493e66517b2861f60738c24f9afa870d90cb5a3b19e0 WatchSource:0}: Error finding container a15d77b97c9b3f13f9ce493e66517b2861f60738c24f9afa870d90cb5a3b19e0: Status 404 returned error can't find the container with id a15d77b97c9b3f13f9ce493e66517b2861f60738c24f9afa870d90cb5a3b19e0 Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.744400 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.765405 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.783047 4884 request.go:700] Waited for 1.92423374s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dmachine-config-server-dockercfg-qx5rd&limit=500&resourceVersion=0 Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.784888 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.825295 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z495x\" (UniqueName: \"kubernetes.io/projected/0b33acfc-c22d-4d1f-a70f-410ed90474bb-kube-api-access-z495x\") pod \"oauth-openshift-558db77b4-wpxzm\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.857510 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c82x\" (UniqueName: \"kubernetes.io/projected/3e56bc10-6d63-4b7b-a740-b449267b3c20-kube-api-access-4c82x\") pod \"cluster-image-registry-operator-dc59b4c8b-fjz74\" (UID: \"3e56bc10-6d63-4b7b-a740-b449267b3c20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.872199 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc6ht\" (UniqueName: \"kubernetes.io/projected/85a7c94e-bc28-4d2e-ad77-7f776252b482-kube-api-access-tc6ht\") pod \"openshift-controller-manager-operator-756b6f6bc6-wthrm\" (UID: \"85a7c94e-bc28-4d2e-ad77-7f776252b482\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.887453 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96x9n\" (UniqueName: \"kubernetes.io/projected/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-kube-api-access-96x9n\") pod \"console-f9d7485db-mh86q\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.897710 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj5lw\" (UniqueName: \"kubernetes.io/projected/3b0da047-6c65-43c9-b8c9-5fef3c6a7929-kube-api-access-rj5lw\") pod \"openshift-config-operator-7777fb866f-bbcm8\" (UID: \"3b0da047-6c65-43c9-b8c9-5fef3c6a7929\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.917160 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.919869 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2m45\" (UniqueName: \"kubernetes.io/projected/5a019c2c-939e-4d88-b662-79dd417dce18-kube-api-access-w2m45\") pod \"apiserver-76f77b778f-sfn66\" (UID: \"5a019c2c-939e-4d88-b662-79dd417dce18\") " pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.939457 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzv9t\" (UniqueName: \"kubernetes.io/projected/5ef99886-22f3-4634-9715-0dd2ee3c53a7-kube-api-access-mzv9t\") pod \"console-operator-58897d9998-vwnln\" (UID: \"5ef99886-22f3-4634-9715-0dd2ee3c53a7\") " pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.941428 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.948030 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.959458 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhs8r\" (UniqueName: \"kubernetes.io/projected/22ccfdc9-0d00-475d-ac54-d084beaca9dd-kube-api-access-jhs8r\") pod \"downloads-7954f5f757-pmdpb\" (UID: \"22ccfdc9-0d00-475d-ac54-d084beaca9dd\") " pod="openshift-console/downloads-7954f5f757-pmdpb" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.963932 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm" Dec 02 01:38:41 crc kubenswrapper[4884]: I1202 01:38:41.978779 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts2rz\" (UniqueName: \"kubernetes.io/projected/d62437d3-3b65-4029-9f01-f6a696e40895-kube-api-access-ts2rz\") pod \"cluster-samples-operator-665b6dd947-n7tvs\" (UID: \"d62437d3-3b65-4029-9f01-f6a696e40895\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n7tvs" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.003267 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46zml\" (UniqueName: \"kubernetes.io/projected/ccc071ec-9724-41f2-a654-625aae8bd840-kube-api-access-46zml\") pod \"controller-manager-879f6c89f-n757x\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.024263 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tnh7\" (UniqueName: \"kubernetes.io/projected/15377bb0-a5a9-4a17-bd32-32795886deac-kube-api-access-9tnh7\") pod \"dns-operator-744455d44c-tk4v5\" (UID: \"15377bb0-a5a9-4a17-bd32-32795886deac\") " pod="openshift-dns-operator/dns-operator-744455d44c-tk4v5" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.047691 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3e56bc10-6d63-4b7b-a740-b449267b3c20-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-fjz74\" (UID: \"3e56bc10-6d63-4b7b-a740-b449267b3c20\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.066248 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frqw9\" (UniqueName: \"kubernetes.io/projected/4517cacd-a375-4c43-a8b5-7d5bf35b172a-kube-api-access-frqw9\") pod \"migrator-59844c95c7-6jbrv\" (UID: \"4517cacd-a375-4c43-a8b5-7d5bf35b172a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6jbrv" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.089232 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmch8\" (UniqueName: \"kubernetes.io/projected/4b95cf91-024d-4571-a28f-b633385428c2-kube-api-access-vmch8\") pod \"kube-storage-version-migrator-operator-b67b599dd-v74zl\" (UID: \"4b95cf91-024d-4571-a28f-b633385428c2\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.095007 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6jbrv" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.103673 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbrq7\" (UniqueName: \"kubernetes.io/projected/a1b20a8b-1487-4658-b022-1112ff5bbe8f-kube-api-access-zbrq7\") pod \"catalog-operator-68c6474976-2txch\" (UID: \"a1b20a8b-1487-4658-b022-1112ff5bbe8f\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.129121 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwt4x\" (UniqueName: \"kubernetes.io/projected/ec5d1cfb-ed76-42f0-a140-447dd49aadb4-kube-api-access-jwt4x\") pod \"packageserver-d55dfcdfc-bfc6m\" (UID: \"ec5d1cfb-ed76-42f0-a140-447dd49aadb4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.140875 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77cbq\" (UniqueName: \"kubernetes.io/projected/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-kube-api-access-77cbq\") pod \"collect-profiles-29410650-wwcg4\" (UID: \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.157148 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.169687 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlw5j\" (UniqueName: \"kubernetes.io/projected/bcb5ff5a-66f7-4bd1-9827-1bb544146761-kube-api-access-zlw5j\") pod \"multus-admission-controller-857f4d67dd-l9nh4\" (UID: \"bcb5ff5a-66f7-4bd1-9827-1bb544146761\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-l9nh4" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.185180 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n7tvs" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.185382 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e6bb8642-f0de-4683-acb3-a50a7492b9a8-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6cxpz\" (UID: \"e6bb8642-f0de-4683-acb3-a50a7492b9a8\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.188486 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mh86q"] Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.200076 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.205580 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbmvd\" (UniqueName: \"kubernetes.io/projected/12396794-f15b-4c87-871b-55094944deb0-kube-api-access-pbmvd\") pod \"machine-config-controller-84d6567774-rwvbr\" (UID: \"12396794-f15b-4c87-871b-55094944deb0\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" Dec 02 01:38:42 crc kubenswrapper[4884]: W1202 01:38:42.205826 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3232fbf_e182_4e1b_a27d_ae2cf61b40b7.slice/crio-82bd2eecf46017729aa2de6ec2c9d8acb3f1f8c4140f498ecab0b728cc87e427 WatchSource:0}: Error finding container 82bd2eecf46017729aa2de6ec2c9d8acb3f1f8c4140f498ecab0b728cc87e427: Status 404 returned error can't find the container with id 82bd2eecf46017729aa2de6ec2c9d8acb3f1f8c4140f498ecab0b728cc87e427 Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.206962 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.220559 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xrls\" (UniqueName: \"kubernetes.io/projected/d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e-kube-api-access-7xrls\") pod \"service-ca-operator-777779d784-n58xw\" (UID: \"d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-n58xw" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.225330 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-pmdpb" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.232422 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-tk4v5" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.240347 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmggs\" (UniqueName: \"kubernetes.io/projected/52dd5b9f-cc6f-4898-a5dd-a235270c39e0-kube-api-access-mmggs\") pod \"router-default-5444994796-664p9\" (UID: \"52dd5b9f-cc6f-4898-a5dd-a235270c39e0\") " pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.255490 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.260097 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlgcn\" (UniqueName: \"kubernetes.io/projected/3c42e876-8a3c-40ce-afc5-eb1599c6edd1-kube-api-access-rlgcn\") pod \"olm-operator-6b444d44fb-4j9b6\" (UID: \"3c42e876-8a3c-40ce-afc5-eb1599c6edd1\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.281119 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kzkw\" (UniqueName: \"kubernetes.io/projected/7bb99172-24fe-4dce-8334-e01d2ff734a2-kube-api-access-9kzkw\") pod \"ingress-operator-5b745b69d9-nlsd7\" (UID: \"7bb99172-24fe-4dce-8334-e01d2ff734a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.308202 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.316839 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7bb99172-24fe-4dce-8334-e01d2ff734a2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-nlsd7\" (UID: \"7bb99172-24fe-4dce-8334-e01d2ff734a2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.317559 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsxpp\" (UniqueName: \"kubernetes.io/projected/db61ff51-5c14-43bd-9042-cf7447ab7a3d-kube-api-access-tsxpp\") pod \"marketplace-operator-79b997595-rdd9f\" (UID: \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.324104 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.329828 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.336604 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.345507 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxpp8\" (UniqueName: \"kubernetes.io/projected/b5c61773-ee25-4653-abdd-304afce58b14-kube-api-access-wxpp8\") pod \"machine-config-operator-74547568cd-fjnvq\" (UID: \"b5c61773-ee25-4653-abdd-304afce58b14\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.359686 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n58xw" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.365512 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.372260 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fncgp\" (UniqueName: \"kubernetes.io/projected/c7c49124-758b-4fab-8567-7bd70b30feb5-kube-api-access-fncgp\") pod \"package-server-manager-789f6589d5-wkddb\" (UID: \"c7c49124-758b-4fab-8567-7bd70b30feb5\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.372382 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.381246 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.391270 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/02d20ecf-4ea9-47e2-b339-7e75e0b8e358-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-fr7hv\" (UID: \"02d20ecf-4ea9-47e2-b339-7e75e0b8e358\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.391786 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.404459 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.424274 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.433598 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm"] Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.436674 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.443290 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-l9nh4" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.444909 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/00e5eb7f-883e-47d3-b342-53401883df9b-etcd-client\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.444951 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e17db01a-f964-4380-ac3f-bcad065313e9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.444974 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/291e4c3e-fffd-4944-b4e7-799fe66c8bbb-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rpqvn\" (UID: \"291e4c3e-fffd-4944-b4e7-799fe66c8bbb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.445055 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-bound-sa-token\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.445078 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68qjq\" (UniqueName: \"kubernetes.io/projected/bddc8712-e48d-4701-aa8a-3cdc96fb7f5a-kube-api-access-68qjq\") pod \"service-ca-9c57cc56f-gkb4d\" (UID: \"bddc8712-e48d-4701-aa8a-3cdc96fb7f5a\") " pod="openshift-service-ca/service-ca-9c57cc56f-gkb4d" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.445110 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6r6gk\" (UniqueName: \"kubernetes.io/projected/00e5eb7f-883e-47d3-b342-53401883df9b-kube-api-access-6r6gk\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.445170 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-registry-tls\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.445185 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00e5eb7f-883e-47d3-b342-53401883df9b-serving-cert\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.445219 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jr49c\" (UniqueName: \"kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-kube-api-access-jr49c\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.445238 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e5eb7f-883e-47d3-b342-53401883df9b-config\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.445271 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/bddc8712-e48d-4701-aa8a-3cdc96fb7f5a-signing-key\") pod \"service-ca-9c57cc56f-gkb4d\" (UID: \"bddc8712-e48d-4701-aa8a-3cdc96fb7f5a\") " pod="openshift-service-ca/service-ca-9c57cc56f-gkb4d" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.446096 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/bddc8712-e48d-4701-aa8a-3cdc96fb7f5a-signing-cabundle\") pod \"service-ca-9c57cc56f-gkb4d\" (UID: \"bddc8712-e48d-4701-aa8a-3cdc96fb7f5a\") " pod="openshift-service-ca/service-ca-9c57cc56f-gkb4d" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.446190 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: E1202 01:38:42.447649 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:42.947626643 +0000 UTC m=+139.623463527 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.450050 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/291e4c3e-fffd-4944-b4e7-799fe66c8bbb-config\") pod \"kube-controller-manager-operator-78b949d7b-rpqvn\" (UID: \"291e4c3e-fffd-4944-b4e7-799fe66c8bbb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.450092 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/00e5eb7f-883e-47d3-b342-53401883df9b-etcd-service-ca\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.450112 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/00e5eb7f-883e-47d3-b342-53401883df9b-etcd-ca\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.450130 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e17db01a-f964-4380-ac3f-bcad065313e9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.450177 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plxbq\" (UniqueName: \"kubernetes.io/projected/e48fed54-40c5-40c5-a88a-71662222bb9e-kube-api-access-plxbq\") pod \"control-plane-machine-set-operator-78cbb6b69f-zq7lm\" (UID: \"e48fed54-40c5-40c5-a88a-71662222bb9e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zq7lm" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.450199 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e17db01a-f964-4380-ac3f-bcad065313e9-registry-certificates\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.450216 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/291e4c3e-fffd-4944-b4e7-799fe66c8bbb-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rpqvn\" (UID: \"291e4c3e-fffd-4944-b4e7-799fe66c8bbb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.450242 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e17db01a-f964-4380-ac3f-bcad065313e9-trusted-ca\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.450257 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e48fed54-40c5-40c5-a88a-71662222bb9e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zq7lm\" (UID: \"e48fed54-40c5-40c5-a88a-71662222bb9e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zq7lm" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.466796 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8"] Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.485361 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wpxzm"] Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.552337 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.552588 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2a972e37-d174-4df4-b278-a3a3c97d0567-cert\") pod \"ingress-canary-jsclc\" (UID: \"2a972e37-d174-4df4-b278-a3a3c97d0567\") " pod="openshift-ingress-canary/ingress-canary-jsclc" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.552644 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p79qw\" (UniqueName: \"kubernetes.io/projected/2a972e37-d174-4df4-b278-a3a3c97d0567-kube-api-access-p79qw\") pod \"ingress-canary-jsclc\" (UID: \"2a972e37-d174-4df4-b278-a3a3c97d0567\") " pod="openshift-ingress-canary/ingress-canary-jsclc" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.552686 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/bb7500d3-5d34-451d-9b19-75ae46e5bf97-socket-dir\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.552736 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgm5x\" (UniqueName: \"kubernetes.io/projected/bb7500d3-5d34-451d-9b19-75ae46e5bf97-kube-api-access-hgm5x\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.552806 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/bddc8712-e48d-4701-aa8a-3cdc96fb7f5a-signing-cabundle\") pod \"service-ca-9c57cc56f-gkb4d\" (UID: \"bddc8712-e48d-4701-aa8a-3cdc96fb7f5a\") " pod="openshift-service-ca/service-ca-9c57cc56f-gkb4d" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.552836 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/bb7500d3-5d34-451d-9b19-75ae46e5bf97-plugins-dir\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.552902 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/291e4c3e-fffd-4944-b4e7-799fe66c8bbb-config\") pod \"kube-controller-manager-operator-78b949d7b-rpqvn\" (UID: \"291e4c3e-fffd-4944-b4e7-799fe66c8bbb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.552958 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/00e5eb7f-883e-47d3-b342-53401883df9b-etcd-service-ca\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.552991 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/00e5eb7f-883e-47d3-b342-53401883df9b-etcd-ca\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553024 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e17db01a-f964-4380-ac3f-bcad065313e9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553041 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/10457084-7272-473a-ab5d-77e2d561e575-node-bootstrap-token\") pod \"machine-config-server-kk9lt\" (UID: \"10457084-7272-473a-ab5d-77e2d561e575\") " pod="openshift-machine-config-operator/machine-config-server-kk9lt" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553087 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plxbq\" (UniqueName: \"kubernetes.io/projected/e48fed54-40c5-40c5-a88a-71662222bb9e-kube-api-access-plxbq\") pod \"control-plane-machine-set-operator-78cbb6b69f-zq7lm\" (UID: \"e48fed54-40c5-40c5-a88a-71662222bb9e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zq7lm" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553130 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e17db01a-f964-4380-ac3f-bcad065313e9-registry-certificates\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553146 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/291e4c3e-fffd-4944-b4e7-799fe66c8bbb-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rpqvn\" (UID: \"291e4c3e-fffd-4944-b4e7-799fe66c8bbb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553162 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/10457084-7272-473a-ab5d-77e2d561e575-certs\") pod \"machine-config-server-kk9lt\" (UID: \"10457084-7272-473a-ab5d-77e2d561e575\") " pod="openshift-machine-config-operator/machine-config-server-kk9lt" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553176 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/bb7500d3-5d34-451d-9b19-75ae46e5bf97-registration-dir\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553207 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e17db01a-f964-4380-ac3f-bcad065313e9-trusted-ca\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553223 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e48fed54-40c5-40c5-a88a-71662222bb9e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zq7lm\" (UID: \"e48fed54-40c5-40c5-a88a-71662222bb9e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zq7lm" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553265 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whxgz\" (UniqueName: \"kubernetes.io/projected/10457084-7272-473a-ab5d-77e2d561e575-kube-api-access-whxgz\") pod \"machine-config-server-kk9lt\" (UID: \"10457084-7272-473a-ab5d-77e2d561e575\") " pod="openshift-machine-config-operator/machine-config-server-kk9lt" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553282 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/00e5eb7f-883e-47d3-b342-53401883df9b-etcd-client\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553397 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e17db01a-f964-4380-ac3f-bcad065313e9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553414 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/291e4c3e-fffd-4944-b4e7-799fe66c8bbb-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rpqvn\" (UID: \"291e4c3e-fffd-4944-b4e7-799fe66c8bbb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553441 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-bound-sa-token\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553457 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68qjq\" (UniqueName: \"kubernetes.io/projected/bddc8712-e48d-4701-aa8a-3cdc96fb7f5a-kube-api-access-68qjq\") pod \"service-ca-9c57cc56f-gkb4d\" (UID: \"bddc8712-e48d-4701-aa8a-3cdc96fb7f5a\") " pod="openshift-service-ca/service-ca-9c57cc56f-gkb4d" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553480 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/bb7500d3-5d34-451d-9b19-75ae46e5bf97-mountpoint-dir\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553515 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6r6gk\" (UniqueName: \"kubernetes.io/projected/00e5eb7f-883e-47d3-b342-53401883df9b-kube-api-access-6r6gk\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553558 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7f827508-03fa-49b2-b625-27235d6b0318-metrics-tls\") pod \"dns-default-74x48\" (UID: \"7f827508-03fa-49b2-b625-27235d6b0318\") " pod="openshift-dns/dns-default-74x48" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553593 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f827508-03fa-49b2-b625-27235d6b0318-config-volume\") pod \"dns-default-74x48\" (UID: \"7f827508-03fa-49b2-b625-27235d6b0318\") " pod="openshift-dns/dns-default-74x48" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553801 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-registry-tls\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553833 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00e5eb7f-883e-47d3-b342-53401883df9b-serving-cert\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553897 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jr49c\" (UniqueName: \"kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-kube-api-access-jr49c\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553933 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e5eb7f-883e-47d3-b342-53401883df9b-config\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553969 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/bb7500d3-5d34-451d-9b19-75ae46e5bf97-csi-data-dir\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.553987 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nwxn\" (UniqueName: \"kubernetes.io/projected/7f827508-03fa-49b2-b625-27235d6b0318-kube-api-access-5nwxn\") pod \"dns-default-74x48\" (UID: \"7f827508-03fa-49b2-b625-27235d6b0318\") " pod="openshift-dns/dns-default-74x48" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.554059 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/bddc8712-e48d-4701-aa8a-3cdc96fb7f5a-signing-key\") pod \"service-ca-9c57cc56f-gkb4d\" (UID: \"bddc8712-e48d-4701-aa8a-3cdc96fb7f5a\") " pod="openshift-service-ca/service-ca-9c57cc56f-gkb4d" Dec 02 01:38:42 crc kubenswrapper[4884]: E1202 01:38:42.554651 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:43.054634545 +0000 UTC m=+139.730471429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.555075 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/291e4c3e-fffd-4944-b4e7-799fe66c8bbb-config\") pod \"kube-controller-manager-operator-78b949d7b-rpqvn\" (UID: \"291e4c3e-fffd-4944-b4e7-799fe66c8bbb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.556436 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/00e5eb7f-883e-47d3-b342-53401883df9b-etcd-service-ca\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.559614 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/bddc8712-e48d-4701-aa8a-3cdc96fb7f5a-signing-cabundle\") pod \"service-ca-9c57cc56f-gkb4d\" (UID: \"bddc8712-e48d-4701-aa8a-3cdc96fb7f5a\") " pod="openshift-service-ca/service-ca-9c57cc56f-gkb4d" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.560495 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00e5eb7f-883e-47d3-b342-53401883df9b-config\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.561596 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e17db01a-f964-4380-ac3f-bcad065313e9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.563520 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/00e5eb7f-883e-47d3-b342-53401883df9b-etcd-ca\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.564825 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e17db01a-f964-4380-ac3f-bcad065313e9-registry-certificates\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.566577 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e17db01a-f964-4380-ac3f-bcad065313e9-trusted-ca\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.580328 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e48fed54-40c5-40c5-a88a-71662222bb9e-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zq7lm\" (UID: \"e48fed54-40c5-40c5-a88a-71662222bb9e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zq7lm" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.588369 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/bddc8712-e48d-4701-aa8a-3cdc96fb7f5a-signing-key\") pod \"service-ca-9c57cc56f-gkb4d\" (UID: \"bddc8712-e48d-4701-aa8a-3cdc96fb7f5a\") " pod="openshift-service-ca/service-ca-9c57cc56f-gkb4d" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.593234 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/291e4c3e-fffd-4944-b4e7-799fe66c8bbb-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rpqvn\" (UID: \"291e4c3e-fffd-4944-b4e7-799fe66c8bbb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.594293 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/00e5eb7f-883e-47d3-b342-53401883df9b-etcd-client\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.594772 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00e5eb7f-883e-47d3-b342-53401883df9b-serving-cert\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.596291 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e17db01a-f964-4380-ac3f-bcad065313e9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.596633 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-registry-tls\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.599713 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6r6gk\" (UniqueName: \"kubernetes.io/projected/00e5eb7f-883e-47d3-b342-53401883df9b-kube-api-access-6r6gk\") pod \"etcd-operator-b45778765-pcf5t\" (UID: \"00e5eb7f-883e-47d3-b342-53401883df9b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.618272 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/291e4c3e-fffd-4944-b4e7-799fe66c8bbb-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rpqvn\" (UID: \"291e4c3e-fffd-4944-b4e7-799fe66c8bbb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.627477 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68qjq\" (UniqueName: \"kubernetes.io/projected/bddc8712-e48d-4701-aa8a-3cdc96fb7f5a-kube-api-access-68qjq\") pod \"service-ca-9c57cc56f-gkb4d\" (UID: \"bddc8712-e48d-4701-aa8a-3cdc96fb7f5a\") " pod="openshift-service-ca/service-ca-9c57cc56f-gkb4d" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.627770 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-bound-sa-token\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.636873 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-664p9" event={"ID":"52dd5b9f-cc6f-4898-a5dd-a235270c39e0","Type":"ContainerStarted","Data":"431c8d50390404ab2ce91b008090bb0ea7f0ba02cea7e84eec4076be4b6657cd"} Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.642465 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" event={"ID":"6b18452c-12ba-4d43-b858-4608001e76d7","Type":"ContainerStarted","Data":"6de95dd0e1c7fe6a880396e332448429b322ee2cc320772fdd981cabf1f9f81f"} Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.642664 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" event={"ID":"6b18452c-12ba-4d43-b858-4608001e76d7","Type":"ContainerStarted","Data":"14b6f527cc3b70a8447efb0bfe4f169b8857578e7b4d9ee5af3d999af81867d8"} Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.642946 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.649696 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.654053 4884 generic.go:334] "Generic (PLEG): container finished" podID="313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed" containerID="457f1d9a27324600706f76942d54114d40fed408693bac966ba37d3d17cefce0" exitCode=0 Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.654178 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" event={"ID":"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed","Type":"ContainerDied","Data":"457f1d9a27324600706f76942d54114d40fed408693bac966ba37d3d17cefce0"} Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.654323 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" event={"ID":"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed","Type":"ContainerStarted","Data":"a15d77b97c9b3f13f9ce493e66517b2861f60738c24f9afa870d90cb5a3b19e0"} Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655105 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/10457084-7272-473a-ab5d-77e2d561e575-certs\") pod \"machine-config-server-kk9lt\" (UID: \"10457084-7272-473a-ab5d-77e2d561e575\") " pod="openshift-machine-config-operator/machine-config-server-kk9lt" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655150 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/bb7500d3-5d34-451d-9b19-75ae46e5bf97-registration-dir\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655189 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whxgz\" (UniqueName: \"kubernetes.io/projected/10457084-7272-473a-ab5d-77e2d561e575-kube-api-access-whxgz\") pod \"machine-config-server-kk9lt\" (UID: \"10457084-7272-473a-ab5d-77e2d561e575\") " pod="openshift-machine-config-operator/machine-config-server-kk9lt" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655225 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/bb7500d3-5d34-451d-9b19-75ae46e5bf97-mountpoint-dir\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655255 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7f827508-03fa-49b2-b625-27235d6b0318-metrics-tls\") pod \"dns-default-74x48\" (UID: \"7f827508-03fa-49b2-b625-27235d6b0318\") " pod="openshift-dns/dns-default-74x48" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655278 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f827508-03fa-49b2-b625-27235d6b0318-config-volume\") pod \"dns-default-74x48\" (UID: \"7f827508-03fa-49b2-b625-27235d6b0318\") " pod="openshift-dns/dns-default-74x48" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655338 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/bb7500d3-5d34-451d-9b19-75ae46e5bf97-csi-data-dir\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655364 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nwxn\" (UniqueName: \"kubernetes.io/projected/7f827508-03fa-49b2-b625-27235d6b0318-kube-api-access-5nwxn\") pod \"dns-default-74x48\" (UID: \"7f827508-03fa-49b2-b625-27235d6b0318\") " pod="openshift-dns/dns-default-74x48" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655397 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2a972e37-d174-4df4-b278-a3a3c97d0567-cert\") pod \"ingress-canary-jsclc\" (UID: \"2a972e37-d174-4df4-b278-a3a3c97d0567\") " pod="openshift-ingress-canary/ingress-canary-jsclc" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655421 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p79qw\" (UniqueName: \"kubernetes.io/projected/2a972e37-d174-4df4-b278-a3a3c97d0567-kube-api-access-p79qw\") pod \"ingress-canary-jsclc\" (UID: \"2a972e37-d174-4df4-b278-a3a3c97d0567\") " pod="openshift-ingress-canary/ingress-canary-jsclc" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655453 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/bb7500d3-5d34-451d-9b19-75ae46e5bf97-socket-dir\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655477 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgm5x\" (UniqueName: \"kubernetes.io/projected/bb7500d3-5d34-451d-9b19-75ae46e5bf97-kube-api-access-hgm5x\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655504 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/bb7500d3-5d34-451d-9b19-75ae46e5bf97-plugins-dir\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655531 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655562 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/10457084-7272-473a-ab5d-77e2d561e575-node-bootstrap-token\") pod \"machine-config-server-kk9lt\" (UID: \"10457084-7272-473a-ab5d-77e2d561e575\") " pod="openshift-machine-config-operator/machine-config-server-kk9lt" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.655736 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/bb7500d3-5d34-451d-9b19-75ae46e5bf97-csi-data-dir\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.656285 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/bb7500d3-5d34-451d-9b19-75ae46e5bf97-mountpoint-dir\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.658041 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/bb7500d3-5d34-451d-9b19-75ae46e5bf97-socket-dir\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.658384 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/bb7500d3-5d34-451d-9b19-75ae46e5bf97-registration-dir\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.658436 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/bb7500d3-5d34-451d-9b19-75ae46e5bf97-plugins-dir\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: E1202 01:38:42.658819 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:43.158805048 +0000 UTC m=+139.834641932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.659243 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7f827508-03fa-49b2-b625-27235d6b0318-config-volume\") pod \"dns-default-74x48\" (UID: \"7f827508-03fa-49b2-b625-27235d6b0318\") " pod="openshift-dns/dns-default-74x48" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.662905 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2a972e37-d174-4df4-b278-a3a3c97d0567-cert\") pod \"ingress-canary-jsclc\" (UID: \"2a972e37-d174-4df4-b278-a3a3c97d0567\") " pod="openshift-ingress-canary/ingress-canary-jsclc" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.663765 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/10457084-7272-473a-ab5d-77e2d561e575-certs\") pod \"machine-config-server-kk9lt\" (UID: \"10457084-7272-473a-ab5d-77e2d561e575\") " pod="openshift-machine-config-operator/machine-config-server-kk9lt" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.664659 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plxbq\" (UniqueName: \"kubernetes.io/projected/e48fed54-40c5-40c5-a88a-71662222bb9e-kube-api-access-plxbq\") pod \"control-plane-machine-set-operator-78cbb6b69f-zq7lm\" (UID: \"e48fed54-40c5-40c5-a88a-71662222bb9e\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zq7lm" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.667568 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/10457084-7272-473a-ab5d-77e2d561e575-node-bootstrap-token\") pod \"machine-config-server-kk9lt\" (UID: \"10457084-7272-473a-ab5d-77e2d561e575\") " pod="openshift-machine-config-operator/machine-config-server-kk9lt" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.668484 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" event={"ID":"2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4","Type":"ContainerStarted","Data":"08a9002c163f88806430f4a3d95423e93e12f3eebfaac13225d9412bb4c9394d"} Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.668548 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" event={"ID":"2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4","Type":"ContainerStarted","Data":"e356b955208e88b1947d86782b8cad5b5addd857f19f8294e68ca08bcf4a6e94"} Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.668564 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" event={"ID":"2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4","Type":"ContainerStarted","Data":"02d235cb9182865ba2833b750b4da501237067e994611b998f841ec8d4aea93b"} Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.676080 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7f827508-03fa-49b2-b625-27235d6b0318-metrics-tls\") pod \"dns-default-74x48\" (UID: \"7f827508-03fa-49b2-b625-27235d6b0318\") " pod="openshift-dns/dns-default-74x48" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.679540 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" event={"ID":"041d3d0d-70e0-4249-a7e2-63958513ba13","Type":"ContainerStarted","Data":"a05f48787a2ffdedc90317f50de38683b6227dbe3bd576473d976566950174a0"} Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.680501 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.684369 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jr49c\" (UniqueName: \"kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-kube-api-access-jr49c\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.696371 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7" event={"ID":"f0d3cbb8-cbf0-4799-b40f-f97beb16a022","Type":"ContainerStarted","Data":"d64b2f3aa021f6f824c898b8cb0aba51396d7ba4dacd3d27d3f1b03fa6a6898f"} Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.704955 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mh86q" event={"ID":"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7","Type":"ContainerStarted","Data":"fe026586176141e5cba9b7b8025e4f8c7038e36d78f7d10f416b794189287a0f"} Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.705001 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mh86q" event={"ID":"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7","Type":"ContainerStarted","Data":"82bd2eecf46017729aa2de6ec2c9d8acb3f1f8c4140f498ecab0b728cc87e427"} Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.712851 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" event={"ID":"d8e078cc-b142-4ccb-b3e1-5c899a354113","Type":"ContainerStarted","Data":"5afb2ddd8aafd973a2f1c3cf912240631e1dab0ff1cabf827baad0d0b187e925"} Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.729409 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whxgz\" (UniqueName: \"kubernetes.io/projected/10457084-7272-473a-ab5d-77e2d561e575-kube-api-access-whxgz\") pod \"machine-config-server-kk9lt\" (UID: \"10457084-7272-473a-ab5d-77e2d561e575\") " pod="openshift-machine-config-operator/machine-config-server-kk9lt" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.729719 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-gkb4d" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.741813 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.750984 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zq7lm" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.754334 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nwxn\" (UniqueName: \"kubernetes.io/projected/7f827508-03fa-49b2-b625-27235d6b0318-kube-api-access-5nwxn\") pod \"dns-default-74x48\" (UID: \"7f827508-03fa-49b2-b625-27235d6b0318\") " pod="openshift-dns/dns-default-74x48" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.761658 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.769931 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p79qw\" (UniqueName: \"kubernetes.io/projected/2a972e37-d174-4df4-b278-a3a3c97d0567-kube-api-access-p79qw\") pod \"ingress-canary-jsclc\" (UID: \"2a972e37-d174-4df4-b278-a3a3c97d0567\") " pod="openshift-ingress-canary/ingress-canary-jsclc" Dec 02 01:38:42 crc kubenswrapper[4884]: E1202 01:38:42.776553 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:43.276532521 +0000 UTC m=+139.952369405 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.783255 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jsclc" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.793347 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-74x48" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.795225 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgm5x\" (UniqueName: \"kubernetes.io/projected/bb7500d3-5d34-451d-9b19-75ae46e5bf97-kube-api-access-hgm5x\") pod \"csi-hostpathplugin-6gvqj\" (UID: \"bb7500d3-5d34-451d-9b19-75ae46e5bf97\") " pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.795913 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-kk9lt" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.852598 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jhnjt" podStartSLOduration=120.85256135 podStartE2EDuration="2m0.85256135s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:42.849921586 +0000 UTC m=+139.525758480" watchObservedRunningTime="2025-12-02 01:38:42.85256135 +0000 UTC m=+139.528398224" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.867564 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:42 crc kubenswrapper[4884]: E1202 01:38:42.867911 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:43.367898503 +0000 UTC m=+140.043735387 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.871662 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.915520 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn" Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.946632 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6jbrv"] Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.964200 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-sfn66"] Dec 02 01:38:42 crc kubenswrapper[4884]: I1202 01:38:42.972537 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:42 crc kubenswrapper[4884]: E1202 01:38:42.973274 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:43.473253314 +0000 UTC m=+140.149090198 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.038924 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" podStartSLOduration=121.038901431 podStartE2EDuration="2m1.038901431s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:43.000888446 +0000 UTC m=+139.676725351" watchObservedRunningTime="2025-12-02 01:38:43.038901431 +0000 UTC m=+139.714738315" Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.082702 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.083171 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" Dec 02 01:38:43 crc kubenswrapper[4884]: E1202 01:38:43.083260 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:43.583239889 +0000 UTC m=+140.259076843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:43 crc kubenswrapper[4884]: W1202 01:38:43.138135 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4517cacd_a375_4c43_a8b5_7d5bf35b172a.slice/crio-c0ab1e9d7ce0a380399d03d74208a8a7a9d70928814443daa78e995342d2f272 WatchSource:0}: Error finding container c0ab1e9d7ce0a380399d03d74208a8a7a9d70928814443daa78e995342d2f272: Status 404 returned error can't find the container with id c0ab1e9d7ce0a380399d03d74208a8a7a9d70928814443daa78e995342d2f272 Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.146095 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-vwnln"] Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.153550 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n7tvs"] Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.183896 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:43 crc kubenswrapper[4884]: E1202 01:38:43.184214 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:43.684069811 +0000 UTC m=+140.359906695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.184347 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:43 crc kubenswrapper[4884]: E1202 01:38:43.184792 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:43.684777488 +0000 UTC m=+140.360614372 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.285301 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:43 crc kubenswrapper[4884]: E1202 01:38:43.285451 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:43.785430806 +0000 UTC m=+140.461267690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.386475 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:43 crc kubenswrapper[4884]: E1202 01:38:43.386866 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:43.886849551 +0000 UTC m=+140.562686435 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.487531 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:43 crc kubenswrapper[4884]: E1202 01:38:43.487729 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:43.987702254 +0000 UTC m=+140.663539158 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.519411 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-4htsc" podStartSLOduration=121.519393985 podStartE2EDuration="2m1.519393985s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:43.515630113 +0000 UTC m=+140.191466997" watchObservedRunningTime="2025-12-02 01:38:43.519393985 +0000 UTC m=+140.195230859" Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.589601 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:43 crc kubenswrapper[4884]: E1202 01:38:43.590041 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:44.090020132 +0000 UTC m=+140.765857026 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.681688 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-mtv5w" podStartSLOduration=121.68166929 podStartE2EDuration="2m1.68166929s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:43.681603208 +0000 UTC m=+140.357440132" watchObservedRunningTime="2025-12-02 01:38:43.68166929 +0000 UTC m=+140.357506184" Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.690360 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:43 crc kubenswrapper[4884]: E1202 01:38:43.690554 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:44.190531465 +0000 UTC m=+140.866368359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.691214 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:43 crc kubenswrapper[4884]: E1202 01:38:43.691637 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:44.191626782 +0000 UTC m=+140.867463676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.733357 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sfn66" event={"ID":"5a019c2c-939e-4d88-b662-79dd417dce18","Type":"ContainerStarted","Data":"e01e56c88d29e8ad1ab73f7d41278258aa1ffc6acd478238b0a92c0cba204b66"} Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.734704 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6jbrv" event={"ID":"4517cacd-a375-4c43-a8b5-7d5bf35b172a","Type":"ContainerStarted","Data":"c0ab1e9d7ce0a380399d03d74208a8a7a9d70928814443daa78e995342d2f272"} Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.736384 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" event={"ID":"3b0da047-6c65-43c9-b8c9-5fef3c6a7929","Type":"ContainerStarted","Data":"002f46a195dba110e33a425a1ae35f469c3e25428bfecf3b4b64b0c118d8a254"} Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.737782 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm" event={"ID":"85a7c94e-bc28-4d2e-ad77-7f776252b482","Type":"ContainerStarted","Data":"e062b5f12f7bfce8e1887a67df8ee8d1004d84148922d392ae60dd58e06ae682"} Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.758881 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-664p9" event={"ID":"52dd5b9f-cc6f-4898-a5dd-a235270c39e0","Type":"ContainerStarted","Data":"36c93106ea1d3e43a95614fa6f9852da07fa2b59b2790b56d1e490a04d617678"} Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.774472 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-kk9lt" event={"ID":"10457084-7272-473a-ab5d-77e2d561e575","Type":"ContainerStarted","Data":"531dfdf5ef70db84cba9683ce4c68838de378e9278c0c8cb2f54a0e6b87ac72f"} Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.777981 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-vwnln" event={"ID":"5ef99886-22f3-4634-9715-0dd2ee3c53a7","Type":"ContainerStarted","Data":"84b54f8a9dce57bc847333d6b1b7590ed5e469d44c58cee7e621f1bcc0a5b141"} Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.782554 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" event={"ID":"0b33acfc-c22d-4d1f-a70f-410ed90474bb","Type":"ContainerStarted","Data":"f23b4b28662d8772cf9741077b471b5d983027cbd37f0e72e22fe60e1be95ba2"} Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.794782 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:43 crc kubenswrapper[4884]: E1202 01:38:43.794895 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:44.294871412 +0000 UTC m=+140.970708296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.795242 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:43 crc kubenswrapper[4884]: E1202 01:38:43.795780 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:44.295736574 +0000 UTC m=+140.971573458 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:43 crc kubenswrapper[4884]: I1202 01:38:43.896517 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:43 crc kubenswrapper[4884]: E1202 01:38:43.898017 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:44.39800028 +0000 UTC m=+141.073837164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:43.999953 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:44 crc kubenswrapper[4884]: E1202 01:38:44.000284 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:44.500272637 +0000 UTC m=+141.176109531 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.101211 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:44 crc kubenswrapper[4884]: E1202 01:38:44.101622 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:44.601606361 +0000 UTC m=+141.277443245 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.202720 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:44 crc kubenswrapper[4884]: E1202 01:38:44.203231 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:44.703217362 +0000 UTC m=+141.379054246 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.234262 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-mh86q" podStartSLOduration=122.234226846 podStartE2EDuration="2m2.234226846s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:44.23315055 +0000 UTC m=+140.908987434" watchObservedRunningTime="2025-12-02 01:38:44.234226846 +0000 UTC m=+140.910063730" Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.305286 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:44 crc kubenswrapper[4884]: E1202 01:38:44.305667 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:44.805650813 +0000 UTC m=+141.481487697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.309486 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.323957 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:44 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:44 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:44 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.324025 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.381932 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-l9nh4"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.395231 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n757x"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.407401 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:44 crc kubenswrapper[4884]: E1202 01:38:44.407783 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:44.907767916 +0000 UTC m=+141.583604800 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.408108 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.410929 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.415229 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-pmdpb"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.421847 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.436801 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.436940 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-tk4v5"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.455556 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.462797 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zq7lm"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.466678 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.502826 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.508025 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-6qvr7" podStartSLOduration=122.508001943 podStartE2EDuration="2m2.508001943s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:44.47988571 +0000 UTC m=+141.155722604" watchObservedRunningTime="2025-12-02 01:38:44.508001943 +0000 UTC m=+141.183838827" Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.508252 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.509145 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:44 crc kubenswrapper[4884]: E1202 01:38:44.509214 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:45.009199673 +0000 UTC m=+141.685036557 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.512811 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:44 crc kubenswrapper[4884]: E1202 01:38:44.513433 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:45.013417255 +0000 UTC m=+141.689254139 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:44 crc kubenswrapper[4884]: W1202 01:38:44.541520 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1b20a8b_1487_4658_b022_1112ff5bbe8f.slice/crio-09866d9bd1ea1120f1291d6a801e060512867f37a2102444d0703d57cf67dbcb WatchSource:0}: Error finding container 09866d9bd1ea1120f1291d6a801e060512867f37a2102444d0703d57cf67dbcb: Status 404 returned error can't find the container with id 09866d9bd1ea1120f1291d6a801e060512867f37a2102444d0703d57cf67dbcb Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.543142 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.551339 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-gkb4d"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.583791 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.601825 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-74x48"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.602689 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.604222 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdd9f"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.622289 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:44 crc kubenswrapper[4884]: E1202 01:38:44.622860 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:45.122820125 +0000 UTC m=+141.798657009 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.658718 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" podStartSLOduration=122.658696168 podStartE2EDuration="2m2.658696168s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:44.644487402 +0000 UTC m=+141.320324286" watchObservedRunningTime="2025-12-02 01:38:44.658696168 +0000 UTC m=+141.334533052" Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.667957 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-pcf5t"] Dec 02 01:38:44 crc kubenswrapper[4884]: W1202 01:38:44.678789 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f827508_03fa_49b2_b625_27235d6b0318.slice/crio-7d3b2535f423108e5734eff65b0b7bde2a56addb06eaca2870ad4aa947556cc5 WatchSource:0}: Error finding container 7d3b2535f423108e5734eff65b0b7bde2a56addb06eaca2870ad4aa947556cc5: Status 404 returned error can't find the container with id 7d3b2535f423108e5734eff65b0b7bde2a56addb06eaca2870ad4aa947556cc5 Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.679128 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jsclc"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.681068 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn"] Dec 02 01:38:44 crc kubenswrapper[4884]: W1202 01:38:44.685882 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00e5eb7f_883e_47d3_b342_53401883df9b.slice/crio-250bbc2cb95ec185cd63742b56960726bbbb4742ff1e6a33d13e297046027374 WatchSource:0}: Error finding container 250bbc2cb95ec185cd63742b56960726bbbb4742ff1e6a33d13e297046027374: Status 404 returned error can't find the container with id 250bbc2cb95ec185cd63742b56960726bbbb4742ff1e6a33d13e297046027374 Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.687643 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-n58xw"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.724553 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:44 crc kubenswrapper[4884]: E1202 01:38:44.724986 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:45.224974669 +0000 UTC m=+141.900811553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.733385 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-664p9" podStartSLOduration=122.733360973 podStartE2EDuration="2m2.733360973s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:44.700624647 +0000 UTC m=+141.376461531" watchObservedRunningTime="2025-12-02 01:38:44.733360973 +0000 UTC m=+141.409197847" Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.753267 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-6gvqj"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.754968 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv"] Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.802959 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-74x48" event={"ID":"7f827508-03fa-49b2-b625-27235d6b0318","Type":"ContainerStarted","Data":"7d3b2535f423108e5734eff65b0b7bde2a56addb06eaca2870ad4aa947556cc5"} Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.804037 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-vwnln" event={"ID":"5ef99886-22f3-4634-9715-0dd2ee3c53a7","Type":"ContainerStarted","Data":"8a14df756100cb1f79da0c6159f8357e0d87b92dd0add22e7f72edf5198b4f22"} Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.805464 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.834479 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:44 crc kubenswrapper[4884]: E1202 01:38:44.835527 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:45.335505577 +0000 UTC m=+142.011342461 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.845292 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zq7lm" event={"ID":"e48fed54-40c5-40c5-a88a-71662222bb9e","Type":"ContainerStarted","Data":"b3f55bf09bfb7ad9778094f9974faaaf81b010b522045606cd762939c7df1844"} Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.859342 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-kk9lt" event={"ID":"10457084-7272-473a-ab5d-77e2d561e575","Type":"ContainerStarted","Data":"e8688c3d036d2f7cb53ea81ca50d6049fa02ae00565facf0739e986e10429f79"} Dec 02 01:38:44 crc kubenswrapper[4884]: W1202 01:38:44.878415 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6d9d6f7_5339_45a4_afd6_2fb4652b3d1e.slice/crio-6e80ed329aca1b754312dfb1dbe1f7764e3708206c6ed3b4f47a35bf039db827 WatchSource:0}: Error finding container 6e80ed329aca1b754312dfb1dbe1f7764e3708206c6ed3b4f47a35bf039db827: Status 404 returned error can't find the container with id 6e80ed329aca1b754312dfb1dbe1f7764e3708206c6ed3b4f47a35bf039db827 Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.879085 4884 generic.go:334] "Generic (PLEG): container finished" podID="5a019c2c-939e-4d88-b662-79dd417dce18" containerID="f25e87de9883bc300c555b5c0941b8ba9b80a452d2dd709805e1ecfcaf3b0f61" exitCode=0 Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.880152 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sfn66" event={"ID":"5a019c2c-939e-4d88-b662-79dd417dce18","Type":"ContainerDied","Data":"f25e87de9883bc300c555b5c0941b8ba9b80a452d2dd709805e1ecfcaf3b0f61"} Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.903261 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb" event={"ID":"c7c49124-758b-4fab-8567-7bd70b30feb5","Type":"ContainerStarted","Data":"cd13397ec3cd4ad23b5ee6ed34435c23e2edf957eb7e9ca5aed48b679a24740c"} Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.916667 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl" event={"ID":"4b95cf91-024d-4571-a28f-b633385428c2","Type":"ContainerStarted","Data":"e9ec8ffa20c7d6b4fc4b01ad789740a90633a3c53b6bb275b27fe5190b575d2e"} Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.942268 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:44 crc kubenswrapper[4884]: E1202 01:38:44.944270 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:45.444256952 +0000 UTC m=+142.120093836 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.949412 4884 generic.go:334] "Generic (PLEG): container finished" podID="3b0da047-6c65-43c9-b8c9-5fef3c6a7929" containerID="8ddce7a3b74222b466acd5e94bb5dab7646db57b617930bcb846d439e0f7ff2b" exitCode=0 Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.949506 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" event={"ID":"3b0da047-6c65-43c9-b8c9-5fef3c6a7929","Type":"ContainerDied","Data":"8ddce7a3b74222b466acd5e94bb5dab7646db57b617930bcb846d439e0f7ff2b"} Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.964348 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz" event={"ID":"e6bb8642-f0de-4683-acb3-a50a7492b9a8","Type":"ContainerStarted","Data":"e6f9497c5587913e72d56e1e7b3632ff8ae47370d0cd860d5027951e76d00cc0"} Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.972471 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-tk4v5" event={"ID":"15377bb0-a5a9-4a17-bd32-32795886deac","Type":"ContainerStarted","Data":"5afe5cc42f98f1e98fddd2c0e9e68e5b56f701241de556db5fa0a6f9ab74bb2a"} Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.982015 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" event={"ID":"b5c61773-ee25-4653-abdd-304afce58b14","Type":"ContainerStarted","Data":"c71b9172566668fe3c8d41b998d4788021b2cf231348a0935a8952c148496d4d"} Dec 02 01:38:44 crc kubenswrapper[4884]: I1202 01:38:44.993153 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" event={"ID":"ec5d1cfb-ed76-42f0-a140-447dd49aadb4","Type":"ContainerStarted","Data":"b60a24d383b8e74330ccb5f9123504aeeb74577dfa9dc8a37dd8dfd705a18b91"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.002782 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" event={"ID":"313b4c60-5ac8-40a8-a2b4-82ab23f5c8ed","Type":"ContainerStarted","Data":"c88fc47ed61ab222ecd7dda7a33eca8b3ba5e9f733076e28f2a2d08af63383c2"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.014350 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-gkb4d" event={"ID":"bddc8712-e48d-4701-aa8a-3cdc96fb7f5a","Type":"ContainerStarted","Data":"5af0dc1ea899647b3d2963a481cc45832f07d300cce53cf9ce4b831911e6ba2b"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.021021 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" event={"ID":"3c42e876-8a3c-40ce-afc5-eb1599c6edd1","Type":"ContainerStarted","Data":"624ef271757951b6a617840cdf7f17bfec51c423c2a53d2cd470dd1ec356b359"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.038717 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" event={"ID":"12396794-f15b-4c87-871b-55094944deb0","Type":"ContainerStarted","Data":"aab320598ee0670d78d7d1b7642d0f48d3efc7b05d482b2b5c9fe9d477c3c00b"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.044274 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.045018 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:45.544992151 +0000 UTC m=+142.220829035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.047375 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" event={"ID":"0b33acfc-c22d-4d1f-a70f-410ed90474bb","Type":"ContainerStarted","Data":"c5b646528bdf8be50f4595e62beed764eaeadefcb1cf90076d48db880c1685b9"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.049101 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.050921 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" event={"ID":"ccc071ec-9724-41f2-a654-625aae8bd840","Type":"ContainerStarted","Data":"1bb44e4774c1bd8bf515059f82c105dfd3ee5eb9d4df49ed51d44409fa2ae6a4"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.062076 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" event={"ID":"00e5eb7f-883e-47d3-b342-53401883df9b","Type":"ContainerStarted","Data":"250bbc2cb95ec185cd63742b56960726bbbb4742ff1e6a33d13e297046027374"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.072505 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" event={"ID":"7bb99172-24fe-4dce-8334-e01d2ff734a2","Type":"ContainerStarted","Data":"d6fa311796e4db15421d685755d0e033888769fa8f7d52dab46c1c789072ece8"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.082188 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-pmdpb" event={"ID":"22ccfdc9-0d00-475d-ac54-d084beaca9dd","Type":"ContainerStarted","Data":"fb65986c07e98bc0f15e40b14fb124852af4184cb47416ab042cfbd81e45a93f"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.095789 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-vwnln" podStartSLOduration=123.095764586 podStartE2EDuration="2m3.095764586s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:45.094002563 +0000 UTC m=+141.769839457" watchObservedRunningTime="2025-12-02 01:38:45.095764586 +0000 UTC m=+141.771601480" Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.098169 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" event={"ID":"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5","Type":"ContainerStarted","Data":"95d7f03a4fda579f3c2b0c64be1a8624875c185a22ce386a96d3773a04c83fbd"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.113061 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-l9nh4" event={"ID":"bcb5ff5a-66f7-4bd1-9827-1bb544146761","Type":"ContainerStarted","Data":"5a251d616f84dcdef94e925fa5473afc099284700af8c86ca98a6f9f4af50a13"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.117878 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" event={"ID":"db61ff51-5c14-43bd-9042-cf7447ab7a3d","Type":"ContainerStarted","Data":"f6711810588bb501331bd68ac442945f4645aa34538d98f345484b9bfdddf167"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.122451 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" event={"ID":"3e56bc10-6d63-4b7b-a740-b449267b3c20","Type":"ContainerStarted","Data":"b2ee9b1868f6473147c75793effa6e53027520458a7230e489a15e03869d7bdb"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.130329 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6jbrv" event={"ID":"4517cacd-a375-4c43-a8b5-7d5bf35b172a","Type":"ContainerStarted","Data":"94c43548c79b1670345b4b5b58ae9073d3546508162ace52e4685ec5afbf48f1"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.137878 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" podStartSLOduration=123.137864709 podStartE2EDuration="2m3.137864709s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:45.137245444 +0000 UTC m=+141.813082348" watchObservedRunningTime="2025-12-02 01:38:45.137864709 +0000 UTC m=+141.813701594" Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.145773 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.146054 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:45.646042328 +0000 UTC m=+142.321879212 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.154880 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.165025 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm" event={"ID":"85a7c94e-bc28-4d2e-ad77-7f776252b482","Type":"ContainerStarted","Data":"bf4a12164a04216da3cc30d575ee60594048658365cfdb337a44a89c98b92c57"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.167386 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn" event={"ID":"291e4c3e-fffd-4944-b4e7-799fe66c8bbb","Type":"ContainerStarted","Data":"a6b768aedaf6b2765ca107d21900e22402a71315c1b5e97586198653da4d71cd"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.172132 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n7tvs" event={"ID":"d62437d3-3b65-4029-9f01-f6a696e40895","Type":"ContainerStarted","Data":"226eb8778f6bae3a0b3f2bd6aa8975c26873c4af8774d83616473b8bd8a46625"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.173838 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" event={"ID":"a1b20a8b-1487-4658-b022-1112ff5bbe8f","Type":"ContainerStarted","Data":"09866d9bd1ea1120f1291d6a801e060512867f37a2102444d0703d57cf67dbcb"} Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.213578 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-kk9lt" podStartSLOduration=6.21356127 podStartE2EDuration="6.21356127s" podCreationTimestamp="2025-12-02 01:38:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:45.211367897 +0000 UTC m=+141.887204781" watchObservedRunningTime="2025-12-02 01:38:45.21356127 +0000 UTC m=+141.889398154" Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.247171 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.247303 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:45.74727634 +0000 UTC m=+142.423113224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.247420 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.248978 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:45.748968151 +0000 UTC m=+142.424805035 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.327102 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:45 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:45 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:45 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.327153 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.336079 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6jbrv" podStartSLOduration=123.335967057 podStartE2EDuration="2m3.335967057s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:45.302832211 +0000 UTC m=+141.978669095" watchObservedRunningTime="2025-12-02 01:38:45.335967057 +0000 UTC m=+142.011803941" Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.348238 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.348430 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:45.848403629 +0000 UTC m=+142.524240513 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.348834 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.349290 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:45.849282831 +0000 UTC m=+142.525119715 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.387600 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" podStartSLOduration=123.387577912 podStartE2EDuration="2m3.387577912s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:45.387239283 +0000 UTC m=+142.063076167" watchObservedRunningTime="2025-12-02 01:38:45.387577912 +0000 UTC m=+142.063414796" Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.391721 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wthrm" podStartSLOduration=123.391703092 podStartE2EDuration="2m3.391703092s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:45.336981791 +0000 UTC m=+142.012818675" watchObservedRunningTime="2025-12-02 01:38:45.391703092 +0000 UTC m=+142.067539996" Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.440237 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-vwnln" Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.452694 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.454984 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:45.95495482 +0000 UTC m=+142.630791704 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.455515 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.456073 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:45.956048277 +0000 UTC m=+142.631885161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.560961 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.561139 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:46.061111941 +0000 UTC m=+142.736948815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.561306 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.561625 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:46.061610174 +0000 UTC m=+142.737447058 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.661737 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.661915 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:46.161891742 +0000 UTC m=+142.837728626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.665925 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.669561 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:46.169536808 +0000 UTC m=+142.845373682 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.769677 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.770586 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:46.270561655 +0000 UTC m=+142.946398539 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.873597 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.874201 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:46.374187214 +0000 UTC m=+143.050024098 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.889453 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.889705 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.945473 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.981730 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.982091 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:46.482064038 +0000 UTC m=+143.157900922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:45 crc kubenswrapper[4884]: I1202 01:38:45.982192 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:45 crc kubenswrapper[4884]: E1202 01:38:45.982615 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:46.482594791 +0000 UTC m=+143.158431675 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.083008 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:46 crc kubenswrapper[4884]: E1202 01:38:46.083400 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:46.583384031 +0000 UTC m=+143.259220915 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.184190 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:46 crc kubenswrapper[4884]: E1202 01:38:46.184810 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:46.684784877 +0000 UTC m=+143.360621761 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.194568 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" event={"ID":"7bb99172-24fe-4dce-8334-e01d2ff734a2","Type":"ContainerStarted","Data":"7824c2bd357e12480245f0bca1146d30329b88e8527b470554a9c1bd27727949"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.197441 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb" event={"ID":"c7c49124-758b-4fab-8567-7bd70b30feb5","Type":"ContainerStarted","Data":"93adeae2c6f94f74cb1026dd2a57b266e88627063c67841af5e7240647f12c8e"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.214976 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n58xw" event={"ID":"d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e","Type":"ContainerStarted","Data":"6e80ed329aca1b754312dfb1dbe1f7764e3708206c6ed3b4f47a35bf039db827"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.218356 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zq7lm" event={"ID":"e48fed54-40c5-40c5-a88a-71662222bb9e","Type":"ContainerStarted","Data":"2db9f860c2104b43371d75bb2d68804059e662c93a62dfe38e4fcfe4f30df0ef"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.250706 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz" event={"ID":"e6bb8642-f0de-4683-acb3-a50a7492b9a8","Type":"ContainerStarted","Data":"6caa77ae313ee9044ab4cb0a3036fffadd661521392b2e2e963c655590b2e0c2"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.278704 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl" event={"ID":"4b95cf91-024d-4571-a28f-b633385428c2","Type":"ContainerStarted","Data":"087cfb54690711ace8dddad58108564a621d4424a8ecb896434207425b4d80ee"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.286300 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:46 crc kubenswrapper[4884]: E1202 01:38:46.286753 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:46.786711276 +0000 UTC m=+143.462548160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.289121 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" event={"ID":"db61ff51-5c14-43bd-9042-cf7447ab7a3d","Type":"ContainerStarted","Data":"eb95cf030ae39b8d31a3de3fe32701ef5083b725345bb12e0d7bab83ddb770e1"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.290245 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.318525 4884 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rdd9f container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.318804 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" podUID="db61ff51-5c14-43bd-9042-cf7447ab7a3d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.319928 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zq7lm" podStartSLOduration=124.319911293 podStartE2EDuration="2m4.319911293s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:46.285840254 +0000 UTC m=+142.961677138" watchObservedRunningTime="2025-12-02 01:38:46.319911293 +0000 UTC m=+142.995748177" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.321678 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-fjz74" event={"ID":"3e56bc10-6d63-4b7b-a740-b449267b3c20","Type":"ContainerStarted","Data":"95621a2ffd9c430ec7fc27b714809600a3cbc542ff3efa36761696012eec1080"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.326864 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:46 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:46 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:46 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.326948 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.337108 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" event={"ID":"ccc071ec-9724-41f2-a654-625aae8bd840","Type":"ContainerStarted","Data":"a62c28c965d6f35f3bdb46093015ed92a0f4a6b9cbfa286adb5dd37ace0a5712"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.337719 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.360448 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.390262 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6cxpz" podStartSLOduration=124.390246573 podStartE2EDuration="2m4.390246573s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:46.319436771 +0000 UTC m=+142.995273655" watchObservedRunningTime="2025-12-02 01:38:46.390246573 +0000 UTC m=+143.066083457" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.390377 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.390578 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-v74zl" podStartSLOduration=124.390572561 podStartE2EDuration="2m4.390572561s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:46.389581957 +0000 UTC m=+143.065418851" watchObservedRunningTime="2025-12-02 01:38:46.390572561 +0000 UTC m=+143.066409445" Dec 02 01:38:46 crc kubenswrapper[4884]: E1202 01:38:46.393630 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:46.893613165 +0000 UTC m=+143.569450039 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.396362 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6jbrv" event={"ID":"4517cacd-a375-4c43-a8b5-7d5bf35b172a","Type":"ContainerStarted","Data":"fd04a15b19ad8993097923fa425147b2aea793d96443fad124b62f989569e41e"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.438559 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n7tvs" event={"ID":"d62437d3-3b65-4029-9f01-f6a696e40895","Type":"ContainerStarted","Data":"0fd50d621c60467480fa68cd31343f0455d9d6943cf26ef7c6939e27aadddfc8"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.464351 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" podStartSLOduration=124.464334765 podStartE2EDuration="2m4.464334765s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:46.463966536 +0000 UTC m=+143.139803420" watchObservedRunningTime="2025-12-02 01:38:46.464334765 +0000 UTC m=+143.140171649" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.486263 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn" event={"ID":"291e4c3e-fffd-4944-b4e7-799fe66c8bbb","Type":"ContainerStarted","Data":"9f4b16e2fbe486b7b04ee0c901adad514eff672c487945a17136c4b489028f75"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.491290 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:46 crc kubenswrapper[4884]: E1202 01:38:46.491659 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:46.991638619 +0000 UTC m=+143.667475503 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.509270 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" event={"ID":"12396794-f15b-4c87-871b-55094944deb0","Type":"ContainerStarted","Data":"290868f9b5cb66e1a123a3ffffb8954d4f681b41da29cc6961d876b326213961"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.535851 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" podStartSLOduration=124.535833243 podStartE2EDuration="2m4.535833243s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:46.535362202 +0000 UTC m=+143.211199086" watchObservedRunningTime="2025-12-02 01:38:46.535833243 +0000 UTC m=+143.211670127" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.551081 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" event={"ID":"b5c61773-ee25-4653-abdd-304afce58b14","Type":"ContainerStarted","Data":"e4331f354466cf1a068895698817749b32d868207d56f4560fb6fd667fd50ff7"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.558478 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-74x48" event={"ID":"7f827508-03fa-49b2-b625-27235d6b0318","Type":"ContainerStarted","Data":"a1ca7daf93b49b36fe124f97172a2801c877c4ab20b2843f7fed3ac104499500"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.559388 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" event={"ID":"a1b20a8b-1487-4658-b022-1112ff5bbe8f","Type":"ContainerStarted","Data":"cfd48669f133d4b71c1ac179c3a4de00023b9b326069a7decf91b0abd4a97221"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.560011 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.592148 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rpqvn" podStartSLOduration=124.592135712 podStartE2EDuration="2m4.592135712s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:46.590081023 +0000 UTC m=+143.265917907" watchObservedRunningTime="2025-12-02 01:38:46.592135712 +0000 UTC m=+143.267972596" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.593398 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:46 crc kubenswrapper[4884]: E1202 01:38:46.594674 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:47.094661434 +0000 UTC m=+143.770498318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.602528 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" event={"ID":"bb7500d3-5d34-451d-9b19-75ae46e5bf97","Type":"ContainerStarted","Data":"d3e042de952470569d572ce064125041da15dedc1acaf325495e8c8dc5ca61c2"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.636913 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-pmdpb" event={"ID":"22ccfdc9-0d00-475d-ac54-d084beaca9dd","Type":"ContainerStarted","Data":"b46ee0ed578869bcd13b68c6d6bb5c8a25aae9f30aa12b2a3f81567d03e3712c"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.637919 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-pmdpb" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.663675 4884 patch_prober.go:28] interesting pod/downloads-7954f5f757-pmdpb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.663779 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-pmdpb" podUID="22ccfdc9-0d00-475d-ac54-d084beaca9dd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.722376 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.724503 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:46 crc kubenswrapper[4884]: E1202 01:38:46.726097 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:47.22607883 +0000 UTC m=+143.901915714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.742368 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-l9nh4" event={"ID":"bcb5ff5a-66f7-4bd1-9827-1bb544146761","Type":"ContainerStarted","Data":"0622e0df1cb5ad71e7d3b8dfe2b715800c2df50e16197f4b922b761878c2b6bc"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.748009 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" podStartSLOduration=124.747998043 podStartE2EDuration="2m4.747998043s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:46.664465661 +0000 UTC m=+143.340302545" watchObservedRunningTime="2025-12-02 01:38:46.747998043 +0000 UTC m=+143.423834927" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.748533 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-2txch" podStartSLOduration=124.748527025 podStartE2EDuration="2m4.748527025s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:46.748139796 +0000 UTC m=+143.423976680" watchObservedRunningTime="2025-12-02 01:38:46.748527025 +0000 UTC m=+143.424363909" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.781583 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" event={"ID":"3c42e876-8a3c-40ce-afc5-eb1599c6edd1","Type":"ContainerStarted","Data":"5f0b527f7a43e521e77276be52c02f5443455fcbcedf128818cda3816b951319"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.782921 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.791026 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jsclc" event={"ID":"2a972e37-d174-4df4-b278-a3a3c97d0567","Type":"ContainerStarted","Data":"94520d1e26909176d1948387de2224ce8730cdbb2d36ad0fad002c8c7843f33a"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.791073 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jsclc" event={"ID":"2a972e37-d174-4df4-b278-a3a3c97d0567","Type":"ContainerStarted","Data":"2e9641b4998d7bf7e4001d0f940d6122da170f7d7d80fbc8f4438792cce2aba3"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.808297 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.818068 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-gkb4d" event={"ID":"bddc8712-e48d-4701-aa8a-3cdc96fb7f5a","Type":"ContainerStarted","Data":"a2886a497902dd1dd14054ab71b6603ea37a8f8b59d38fe2639c018adb614da4"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.828951 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:46 crc kubenswrapper[4884]: E1202 01:38:46.829495 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:47.329482704 +0000 UTC m=+144.005319588 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.830088 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-pmdpb" podStartSLOduration=124.830071488 podStartE2EDuration="2m4.830071488s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:46.829667668 +0000 UTC m=+143.505504542" watchObservedRunningTime="2025-12-02 01:38:46.830071488 +0000 UTC m=+143.505908372" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.862080 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv" event={"ID":"02d20ecf-4ea9-47e2-b339-7e75e0b8e358","Type":"ContainerStarted","Data":"d986f4cfcbcd1c42a6191557fc84d38078747fe253d7d8e52d69568d417cf1d2"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.867076 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-gkb4d" podStartSLOduration=124.867062518 podStartE2EDuration="2m4.867062518s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:46.866212117 +0000 UTC m=+143.542049001" watchObservedRunningTime="2025-12-02 01:38:46.867062518 +0000 UTC m=+143.542899402" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.876568 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" event={"ID":"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5","Type":"ContainerStarted","Data":"1f29e1274403e52b260ee461e939515a522aeb1b9ea0882d6e1b619c53638712"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.899866 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-jsclc" podStartSLOduration=7.899852305 podStartE2EDuration="7.899852305s" podCreationTimestamp="2025-12-02 01:38:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:46.89882323 +0000 UTC m=+143.574660114" watchObservedRunningTime="2025-12-02 01:38:46.899852305 +0000 UTC m=+143.575689189" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.924389 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-tk4v5" event={"ID":"15377bb0-a5a9-4a17-bd32-32795886deac","Type":"ContainerStarted","Data":"c19c4ea364e18c6cafc90c5904d19fba8f61844a4b0ddfaa33561fb32293cdb3"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.929861 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:46 crc kubenswrapper[4884]: E1202 01:38:46.934363 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:47.434344304 +0000 UTC m=+144.110181188 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.959541 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" event={"ID":"ec5d1cfb-ed76-42f0-a140-447dd49aadb4","Type":"ContainerStarted","Data":"e9c184a74da71d8da194cf1c8c3bc42129189ea888bd68b4b92958b5c4fb3ac6"} Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.959600 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.971908 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.971962 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.975545 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-cswlp" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.989175 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv" podStartSLOduration=124.989160847 podStartE2EDuration="2m4.989160847s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:46.987825204 +0000 UTC m=+143.663662088" watchObservedRunningTime="2025-12-02 01:38:46.989160847 +0000 UTC m=+143.664997731" Dec 02 01:38:46 crc kubenswrapper[4884]: I1202 01:38:46.990505 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4j9b6" podStartSLOduration=124.990498409 podStartE2EDuration="2m4.990498409s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:46.940021322 +0000 UTC m=+143.615858206" watchObservedRunningTime="2025-12-02 01:38:46.990498409 +0000 UTC m=+143.666335293" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.021585 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" podStartSLOduration=125.021569405 podStartE2EDuration="2m5.021569405s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:47.021106714 +0000 UTC m=+143.696943598" watchObservedRunningTime="2025-12-02 01:38:47.021569405 +0000 UTC m=+143.697406289" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.031857 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:47 crc kubenswrapper[4884]: E1202 01:38:47.034987 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:47.534973181 +0000 UTC m=+144.210810125 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.055267 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-tk4v5" podStartSLOduration=125.055249564 podStartE2EDuration="2m5.055249564s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:47.055063069 +0000 UTC m=+143.730899963" watchObservedRunningTime="2025-12-02 01:38:47.055249564 +0000 UTC m=+143.731086448" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.137649 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:47 crc kubenswrapper[4884]: E1202 01:38:47.138058 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:47.638041587 +0000 UTC m=+144.313878471 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.151029 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.158582 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-bfc6m" podStartSLOduration=125.158560406 podStartE2EDuration="2m5.158560406s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:47.158142126 +0000 UTC m=+143.833979010" watchObservedRunningTime="2025-12-02 01:38:47.158560406 +0000 UTC m=+143.834397300" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.240481 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:47 crc kubenswrapper[4884]: E1202 01:38:47.240868 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:47.740854746 +0000 UTC m=+144.416691630 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.327007 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:47 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:47 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:47 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.327432 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.332126 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tqtn7"] Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.333077 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.338290 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.342076 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:47 crc kubenswrapper[4884]: E1202 01:38:47.342489 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:47.842476747 +0000 UTC m=+144.518313631 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.356907 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tqtn7"] Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.443497 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg8v8\" (UniqueName: \"kubernetes.io/projected/525da77f-9387-441b-8ded-8f7b1cc41df8-kube-api-access-pg8v8\") pod \"community-operators-tqtn7\" (UID: \"525da77f-9387-441b-8ded-8f7b1cc41df8\") " pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.443723 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/525da77f-9387-441b-8ded-8f7b1cc41df8-utilities\") pod \"community-operators-tqtn7\" (UID: \"525da77f-9387-441b-8ded-8f7b1cc41df8\") " pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.443868 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.443992 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/525da77f-9387-441b-8ded-8f7b1cc41df8-catalog-content\") pod \"community-operators-tqtn7\" (UID: \"525da77f-9387-441b-8ded-8f7b1cc41df8\") " pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:38:47 crc kubenswrapper[4884]: E1202 01:38:47.444360 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:47.944347414 +0000 UTC m=+144.620184298 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.544933 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2mzck"] Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.545319 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:47 crc kubenswrapper[4884]: E1202 01:38:47.545503 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:48.045474433 +0000 UTC m=+144.721311317 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.545619 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/525da77f-9387-441b-8ded-8f7b1cc41df8-catalog-content\") pod \"community-operators-tqtn7\" (UID: \"525da77f-9387-441b-8ded-8f7b1cc41df8\") " pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.545704 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg8v8\" (UniqueName: \"kubernetes.io/projected/525da77f-9387-441b-8ded-8f7b1cc41df8-kube-api-access-pg8v8\") pod \"community-operators-tqtn7\" (UID: \"525da77f-9387-441b-8ded-8f7b1cc41df8\") " pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.545756 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/525da77f-9387-441b-8ded-8f7b1cc41df8-utilities\") pod \"community-operators-tqtn7\" (UID: \"525da77f-9387-441b-8ded-8f7b1cc41df8\") " pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.545805 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:47 crc kubenswrapper[4884]: E1202 01:38:47.546122 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:48.046108049 +0000 UTC m=+144.721944933 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.546475 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/525da77f-9387-441b-8ded-8f7b1cc41df8-catalog-content\") pod \"community-operators-tqtn7\" (UID: \"525da77f-9387-441b-8ded-8f7b1cc41df8\") " pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.546797 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/525da77f-9387-441b-8ded-8f7b1cc41df8-utilities\") pod \"community-operators-tqtn7\" (UID: \"525da77f-9387-441b-8ded-8f7b1cc41df8\") " pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.547180 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.552472 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.556879 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2mzck"] Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.652693 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.653180 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg8v8\" (UniqueName: \"kubernetes.io/projected/525da77f-9387-441b-8ded-8f7b1cc41df8-kube-api-access-pg8v8\") pod \"community-operators-tqtn7\" (UID: \"525da77f-9387-441b-8ded-8f7b1cc41df8\") " pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.653417 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-catalog-content\") pod \"certified-operators-2mzck\" (UID: \"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80\") " pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.653522 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2dmz\" (UniqueName: \"kubernetes.io/projected/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-kube-api-access-m2dmz\") pod \"certified-operators-2mzck\" (UID: \"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80\") " pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.653666 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-utilities\") pod \"certified-operators-2mzck\" (UID: \"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80\") " pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:38:47 crc kubenswrapper[4884]: E1202 01:38:47.653997 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:48.153973672 +0000 UTC m=+144.829810556 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.654400 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.758455 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-utilities\") pod \"certified-operators-2mzck\" (UID: \"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80\") " pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.758978 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.759007 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-catalog-content\") pod \"certified-operators-2mzck\" (UID: \"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80\") " pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.759051 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2dmz\" (UniqueName: \"kubernetes.io/projected/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-kube-api-access-m2dmz\") pod \"certified-operators-2mzck\" (UID: \"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80\") " pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.759251 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-utilities\") pod \"certified-operators-2mzck\" (UID: \"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80\") " pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:38:47 crc kubenswrapper[4884]: E1202 01:38:47.759551 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:48.259538589 +0000 UTC m=+144.935375473 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.759722 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-catalog-content\") pod \"certified-operators-2mzck\" (UID: \"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80\") " pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.811733 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6vm82"] Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.812617 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.855710 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2dmz\" (UniqueName: \"kubernetes.io/projected/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-kube-api-access-m2dmz\") pod \"certified-operators-2mzck\" (UID: \"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80\") " pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.867410 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.867619 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.867695 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76949538-95e6-4f55-a9e3-3dcccd40580e-catalog-content\") pod \"community-operators-6vm82\" (UID: \"76949538-95e6-4f55-a9e3-3dcccd40580e\") " pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.867732 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76949538-95e6-4f55-a9e3-3dcccd40580e-utilities\") pod \"community-operators-6vm82\" (UID: \"76949538-95e6-4f55-a9e3-3dcccd40580e\") " pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.867834 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxzgn\" (UniqueName: \"kubernetes.io/projected/76949538-95e6-4f55-a9e3-3dcccd40580e-kube-api-access-vxzgn\") pod \"community-operators-6vm82\" (UID: \"76949538-95e6-4f55-a9e3-3dcccd40580e\") " pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:38:47 crc kubenswrapper[4884]: E1202 01:38:47.867999 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:48.367978136 +0000 UTC m=+145.043815020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.877569 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6vm82"] Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.976522 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.976598 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76949538-95e6-4f55-a9e3-3dcccd40580e-catalog-content\") pod \"community-operators-6vm82\" (UID: \"76949538-95e6-4f55-a9e3-3dcccd40580e\") " pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.976632 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76949538-95e6-4f55-a9e3-3dcccd40580e-utilities\") pod \"community-operators-6vm82\" (UID: \"76949538-95e6-4f55-a9e3-3dcccd40580e\") " pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.976693 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxzgn\" (UniqueName: \"kubernetes.io/projected/76949538-95e6-4f55-a9e3-3dcccd40580e-kube-api-access-vxzgn\") pod \"community-operators-6vm82\" (UID: \"76949538-95e6-4f55-a9e3-3dcccd40580e\") " pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:38:47 crc kubenswrapper[4884]: E1202 01:38:47.977101 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:48.477082799 +0000 UTC m=+145.152919673 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.977199 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76949538-95e6-4f55-a9e3-3dcccd40580e-catalog-content\") pod \"community-operators-6vm82\" (UID: \"76949538-95e6-4f55-a9e3-3dcccd40580e\") " pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:38:47 crc kubenswrapper[4884]: I1202 01:38:47.977267 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76949538-95e6-4f55-a9e3-3dcccd40580e-utilities\") pod \"community-operators-6vm82\" (UID: \"76949538-95e6-4f55-a9e3-3dcccd40580e\") " pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.005154 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-26msg"] Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.006047 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.038989 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" event={"ID":"7bb99172-24fe-4dce-8334-e01d2ff734a2","Type":"ContainerStarted","Data":"8ffb93a3a81897c4e1da365104fa67abd5373b52e77d64c70baf548202c3a261"} Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.054643 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxzgn\" (UniqueName: \"kubernetes.io/projected/76949538-95e6-4f55-a9e3-3dcccd40580e-kube-api-access-vxzgn\") pod \"community-operators-6vm82\" (UID: \"76949538-95e6-4f55-a9e3-3dcccd40580e\") " pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.067250 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n58xw" event={"ID":"d6d9d6f7-5339-45a4-afd6-2fb4652b3d1e","Type":"ContainerStarted","Data":"b1ee822792f7f388e2874a7c619bb62897f12207f58af82ffafd9a5ea3dd505c"} Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.074213 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-26msg"] Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.080277 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.080483 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39284c28-156f-445d-9951-9df6851c9487-catalog-content\") pod \"certified-operators-26msg\" (UID: \"39284c28-156f-445d-9951-9df6851c9487\") " pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.080526 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39284c28-156f-445d-9951-9df6851c9487-utilities\") pod \"certified-operators-26msg\" (UID: \"39284c28-156f-445d-9951-9df6851c9487\") " pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.080569 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-829qm\" (UniqueName: \"kubernetes.io/projected/39284c28-156f-445d-9951-9df6851c9487-kube-api-access-829qm\") pod \"certified-operators-26msg\" (UID: \"39284c28-156f-445d-9951-9df6851c9487\") " pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:38:48 crc kubenswrapper[4884]: E1202 01:38:48.080684 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:48.580669978 +0000 UTC m=+145.256506852 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.109004 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-fjnvq" event={"ID":"b5c61773-ee25-4653-abdd-304afce58b14","Type":"ContainerStarted","Data":"a5aa962f3627a9bc07d401e5cfeb85c4773b48525f733ab68d9e20ee34a01528"} Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.131532 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.159266 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb" event={"ID":"c7c49124-758b-4fab-8567-7bd70b30feb5","Type":"ContainerStarted","Data":"02f560157df18fe88a5bfc142ff1524ed405ce889ae2506ffd354a0597354f83"} Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.159435 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.183558 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-829qm\" (UniqueName: \"kubernetes.io/projected/39284c28-156f-445d-9951-9df6851c9487-kube-api-access-829qm\") pod \"certified-operators-26msg\" (UID: \"39284c28-156f-445d-9951-9df6851c9487\") " pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.183677 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39284c28-156f-445d-9951-9df6851c9487-catalog-content\") pod \"certified-operators-26msg\" (UID: \"39284c28-156f-445d-9951-9df6851c9487\") " pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.183714 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.183794 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39284c28-156f-445d-9951-9df6851c9487-utilities\") pod \"certified-operators-26msg\" (UID: \"39284c28-156f-445d-9951-9df6851c9487\") " pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.186405 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39284c28-156f-445d-9951-9df6851c9487-catalog-content\") pod \"certified-operators-26msg\" (UID: \"39284c28-156f-445d-9951-9df6851c9487\") " pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:38:48 crc kubenswrapper[4884]: E1202 01:38:48.190162 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:48.69015134 +0000 UTC m=+145.365988224 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.191135 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39284c28-156f-445d-9951-9df6851c9487-utilities\") pod \"certified-operators-26msg\" (UID: \"39284c28-156f-445d-9951-9df6851c9487\") " pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.200717 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-74x48" event={"ID":"7f827508-03fa-49b2-b625-27235d6b0318","Type":"ContainerStarted","Data":"ae0ac6bf31a7abb9660c79a573ca67c126335b662e62312e05408e6c432611d8"} Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.201454 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-74x48" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.206862 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" event={"ID":"3b0da047-6c65-43c9-b8c9-5fef3c6a7929","Type":"ContainerStarted","Data":"34f80a7638ad65d1ca3e7e19633299fed73854aa34ee4598211f36b458db9189"} Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.207335 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.215820 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlsd7" podStartSLOduration=126.215801534 podStartE2EDuration="2m6.215801534s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:48.170208815 +0000 UTC m=+144.846045709" watchObservedRunningTime="2025-12-02 01:38:48.215801534 +0000 UTC m=+144.891638418" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.223226 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n7tvs" event={"ID":"d62437d3-3b65-4029-9f01-f6a696e40895","Type":"ContainerStarted","Data":"b16fbdb753021dce681a8f78dbe406b02afe4549533cd3b091f58f7b2e1bb65c"} Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.246930 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-tk4v5" event={"ID":"15377bb0-a5a9-4a17-bd32-32795886deac","Type":"ContainerStarted","Data":"806f47d75734e5097964534d50463296a0a099b44b9c6c0f9de12a540e9a7325"} Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.248260 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-829qm\" (UniqueName: \"kubernetes.io/projected/39284c28-156f-445d-9951-9df6851c9487-kube-api-access-829qm\") pod \"certified-operators-26msg\" (UID: \"39284c28-156f-445d-9951-9df6851c9487\") " pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.252733 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" event={"ID":"12396794-f15b-4c87-871b-55094944deb0","Type":"ContainerStarted","Data":"0b1b8402a4b642c51fc04665bb476c830b6579788e26f52d761ff83cf96adbe3"} Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.258003 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-n58xw" podStartSLOduration=126.257983639 podStartE2EDuration="2m6.257983639s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:48.21769178 +0000 UTC m=+144.893528664" watchObservedRunningTime="2025-12-02 01:38:48.257983639 +0000 UTC m=+144.933820523" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.260042 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb" podStartSLOduration=126.260034239 podStartE2EDuration="2m6.260034239s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:48.251093472 +0000 UTC m=+144.926930356" watchObservedRunningTime="2025-12-02 01:38:48.260034239 +0000 UTC m=+144.935871123" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.288413 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:48 crc kubenswrapper[4884]: E1202 01:38:48.289274 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:48.789246129 +0000 UTC m=+145.465083013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.301472 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" event={"ID":"bb7500d3-5d34-451d-9b19-75ae46e5bf97","Type":"ContainerStarted","Data":"6c8682e594365685df73cfc950d94df8ff133abac9b466e5667e0281beabb556"} Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.315598 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:48 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:48 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:48 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.315933 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.341339 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" podStartSLOduration=126.341323626 podStartE2EDuration="2m6.341323626s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:48.287812285 +0000 UTC m=+144.963649169" watchObservedRunningTime="2025-12-02 01:38:48.341323626 +0000 UTC m=+145.017160510" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.343088 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-l9nh4" event={"ID":"bcb5ff5a-66f7-4bd1-9827-1bb544146761","Type":"ContainerStarted","Data":"b93a00dffebf0dd2874d237f85dbc7d60478a0a4208fe1b3f294748f093f6abe"} Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.378108 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rwvbr" podStartSLOduration=126.37809259 podStartE2EDuration="2m6.37809259s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:48.374942733 +0000 UTC m=+145.050779607" watchObservedRunningTime="2025-12-02 01:38:48.37809259 +0000 UTC m=+145.053929474" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.378584 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-74x48" podStartSLOduration=9.378578182 podStartE2EDuration="9.378578182s" podCreationTimestamp="2025-12-02 01:38:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:48.342752631 +0000 UTC m=+145.018589535" watchObservedRunningTime="2025-12-02 01:38:48.378578182 +0000 UTC m=+145.054415066" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.392309 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:48 crc kubenswrapper[4884]: E1202 01:38:48.392981 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:48.892968272 +0000 UTC m=+145.568805156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.411697 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sfn66" event={"ID":"5a019c2c-939e-4d88-b662-79dd417dce18","Type":"ContainerStarted","Data":"5af17042d7a3aa639f8a76f3bc7e3e9a69e3b4c4f49245c028839cb4bb006687"} Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.411735 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-sfn66" event={"ID":"5a019c2c-939e-4d88-b662-79dd417dce18","Type":"ContainerStarted","Data":"221adf2d37abc7727cec7ebfe8736b1a55983b5bdfe5035a1477e74bcfedaf4b"} Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.417724 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-n7tvs" podStartSLOduration=126.417710763 podStartE2EDuration="2m6.417710763s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:48.416945675 +0000 UTC m=+145.092782569" watchObservedRunningTime="2025-12-02 01:38:48.417710763 +0000 UTC m=+145.093547647" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.421626 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-fr7hv" event={"ID":"02d20ecf-4ea9-47e2-b339-7e75e0b8e358","Type":"ContainerStarted","Data":"c1171b172189e62f54804129b8161c086ac027451c2d464617f267125cd4e04b"} Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.424917 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" event={"ID":"00e5eb7f-883e-47d3-b342-53401883df9b","Type":"ContainerStarted","Data":"4de51b2b35c09664ac24932869d687f480b7247f2051faa30954f9fe14da5c7c"} Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.431899 4884 patch_prober.go:28] interesting pod/downloads-7954f5f757-pmdpb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.431957 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-pmdpb" podUID="22ccfdc9-0d00-475d-ac54-d084beaca9dd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.434115 4884 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rdd9f container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.434144 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" podUID="db61ff51-5c14-43bd-9042-cf7447ab7a3d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.435407 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.494274 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:48 crc kubenswrapper[4884]: E1202 01:38:48.495961 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:48.995946966 +0000 UTC m=+145.671783850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.550224 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-l9nh4" podStartSLOduration=126.550191245 podStartE2EDuration="2m6.550191245s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:48.490907013 +0000 UTC m=+145.166743897" watchObservedRunningTime="2025-12-02 01:38:48.550191245 +0000 UTC m=+145.226028129" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.601929 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tqtn7"] Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.610502 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-sfn66" podStartSLOduration=126.610482881 podStartE2EDuration="2m6.610482881s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:48.599288089 +0000 UTC m=+145.275124993" watchObservedRunningTime="2025-12-02 01:38:48.610482881 +0000 UTC m=+145.286319765" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.619406 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:48 crc kubenswrapper[4884]: E1202 01:38:48.620019 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:49.120006592 +0000 UTC m=+145.795843476 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.724265 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:48 crc kubenswrapper[4884]: E1202 01:38:48.724593 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:49.224577845 +0000 UTC m=+145.900414729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.791474 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-pcf5t" podStartSLOduration=126.791459162 podStartE2EDuration="2m6.791459162s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:48.673061983 +0000 UTC m=+145.348898867" watchObservedRunningTime="2025-12-02 01:38:48.791459162 +0000 UTC m=+145.467296046" Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.792195 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2mzck"] Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.825505 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:48 crc kubenswrapper[4884]: E1202 01:38:48.838007 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:49.337986583 +0000 UTC m=+146.013823467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.897369 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6vm82"] Dec 02 01:38:48 crc kubenswrapper[4884]: I1202 01:38:48.927662 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:48 crc kubenswrapper[4884]: E1202 01:38:48.928083 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:49.428066233 +0000 UTC m=+146.103903117 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.030961 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:49 crc kubenswrapper[4884]: E1202 01:38:49.031792 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:49.531779095 +0000 UTC m=+146.207615979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.103083 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-26msg"] Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.111595 4884 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.135338 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:49 crc kubenswrapper[4884]: E1202 01:38:49.135731 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:49.635716593 +0000 UTC m=+146.311553477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.237205 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:49 crc kubenswrapper[4884]: E1202 01:38:49.237592 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:49.73757577 +0000 UTC m=+146.413412654 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.314217 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:49 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:49 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:49 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.314314 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.338344 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:49 crc kubenswrapper[4884]: E1202 01:38:49.338566 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:49.838533415 +0000 UTC m=+146.514370299 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.338643 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:49 crc kubenswrapper[4884]: E1202 01:38:49.339084 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:49.839067738 +0000 UTC m=+146.514904612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.430401 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vm82" event={"ID":"76949538-95e6-4f55-a9e3-3dcccd40580e","Type":"ContainerStarted","Data":"47e8c5a549e12fa67e0ed20e228d780aa8d2b6d720c700dbfe5d60e8c49478f4"} Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.437929 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26msg" event={"ID":"39284c28-156f-445d-9951-9df6851c9487","Type":"ContainerStarted","Data":"73bf0058ddf2441a7b2d9a2890ec92b4c6023d9db35c2b2977efb7e0d8301f8a"} Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.439938 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:49 crc kubenswrapper[4884]: E1202 01:38:49.440150 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:49.940120155 +0000 UTC m=+146.615957039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.440294 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:49 crc kubenswrapper[4884]: E1202 01:38:49.440788 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:49.940771111 +0000 UTC m=+146.616607995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.445976 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mzck" event={"ID":"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80","Type":"ContainerStarted","Data":"a1774190940f7913f4531396bd52b132b0f1f6578b889bfed8aa7b93bc56aa12"} Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.446029 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mzck" event={"ID":"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80","Type":"ContainerStarted","Data":"050bbf1f3c94858339b102e0a95d2d4bed2f6bd7429cde8dd796a406c8f59a07"} Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.448164 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" event={"ID":"bb7500d3-5d34-451d-9b19-75ae46e5bf97","Type":"ContainerStarted","Data":"4baa382e3f7e152b408f0bfe0890d471aa2d27cd73d11c502336908e69faa1ae"} Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.450136 4884 generic.go:334] "Generic (PLEG): container finished" podID="525da77f-9387-441b-8ded-8f7b1cc41df8" containerID="5c99f3ee4ed330750d81af7ac344010f66fc61f741191b7320a2a5de40434169" exitCode=0 Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.450426 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqtn7" event={"ID":"525da77f-9387-441b-8ded-8f7b1cc41df8","Type":"ContainerDied","Data":"5c99f3ee4ed330750d81af7ac344010f66fc61f741191b7320a2a5de40434169"} Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.450494 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqtn7" event={"ID":"525da77f-9387-441b-8ded-8f7b1cc41df8","Type":"ContainerStarted","Data":"f879d117e5598a7c9dde048cf366ec86b4ead1ceea2cb77d8478d2b1b7c11fb6"} Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.451464 4884 patch_prober.go:28] interesting pod/downloads-7954f5f757-pmdpb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.451584 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-pmdpb" podUID="22ccfdc9-0d00-475d-ac54-d084beaca9dd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.454939 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.462085 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-bbcm8" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.537180 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k7v6v"] Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.538385 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.543821 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:49 crc kubenswrapper[4884]: E1202 01:38:49.545584 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:50.045566089 +0000 UTC m=+146.721402973 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.561206 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.586056 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7v6v"] Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.651195 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tx4l\" (UniqueName: \"kubernetes.io/projected/25c18b0c-a129-43c2-8d8e-8381d421510a-kube-api-access-7tx4l\") pod \"redhat-marketplace-k7v6v\" (UID: \"25c18b0c-a129-43c2-8d8e-8381d421510a\") " pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.651288 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25c18b0c-a129-43c2-8d8e-8381d421510a-catalog-content\") pod \"redhat-marketplace-k7v6v\" (UID: \"25c18b0c-a129-43c2-8d8e-8381d421510a\") " pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.651340 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.651368 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25c18b0c-a129-43c2-8d8e-8381d421510a-utilities\") pod \"redhat-marketplace-k7v6v\" (UID: \"25c18b0c-a129-43c2-8d8e-8381d421510a\") " pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:38:49 crc kubenswrapper[4884]: E1202 01:38:49.651822 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:50.151806052 +0000 UTC m=+146.827642936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.753835 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.754876 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25c18b0c-a129-43c2-8d8e-8381d421510a-utilities\") pod \"redhat-marketplace-k7v6v\" (UID: \"25c18b0c-a129-43c2-8d8e-8381d421510a\") " pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.754940 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tx4l\" (UniqueName: \"kubernetes.io/projected/25c18b0c-a129-43c2-8d8e-8381d421510a-kube-api-access-7tx4l\") pod \"redhat-marketplace-k7v6v\" (UID: \"25c18b0c-a129-43c2-8d8e-8381d421510a\") " pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.754998 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25c18b0c-a129-43c2-8d8e-8381d421510a-catalog-content\") pod \"redhat-marketplace-k7v6v\" (UID: \"25c18b0c-a129-43c2-8d8e-8381d421510a\") " pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.755558 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25c18b0c-a129-43c2-8d8e-8381d421510a-catalog-content\") pod \"redhat-marketplace-k7v6v\" (UID: \"25c18b0c-a129-43c2-8d8e-8381d421510a\") " pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:38:49 crc kubenswrapper[4884]: E1202 01:38:49.755657 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-02 01:38:50.255635907 +0000 UTC m=+146.931472791 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.755899 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25c18b0c-a129-43c2-8d8e-8381d421510a-utilities\") pod \"redhat-marketplace-k7v6v\" (UID: \"25c18b0c-a129-43c2-8d8e-8381d421510a\") " pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.785770 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tx4l\" (UniqueName: \"kubernetes.io/projected/25c18b0c-a129-43c2-8d8e-8381d421510a-kube-api-access-7tx4l\") pod \"redhat-marketplace-k7v6v\" (UID: \"25c18b0c-a129-43c2-8d8e-8381d421510a\") " pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.811758 4884 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-02T01:38:49.111696829Z","Handler":null,"Name":""} Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.856554 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:49 crc kubenswrapper[4884]: E1202 01:38:49.857058 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-02 01:38:50.357022183 +0000 UTC m=+147.032859067 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-cmk67" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.873247 4884 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.873289 4884 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.919653 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ngw4l"] Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.920613 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.932228 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ngw4l"] Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.950737 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.959171 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.959501 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fbea3c3-376b-48e1-a863-3eb8478a3974-utilities\") pod \"redhat-marketplace-ngw4l\" (UID: \"8fbea3c3-376b-48e1-a863-3eb8478a3974\") " pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.959550 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fbea3c3-376b-48e1-a863-3eb8478a3974-catalog-content\") pod \"redhat-marketplace-ngw4l\" (UID: \"8fbea3c3-376b-48e1-a863-3eb8478a3974\") " pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.959595 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f8x6\" (UniqueName: \"kubernetes.io/projected/8fbea3c3-376b-48e1-a863-3eb8478a3974-kube-api-access-2f8x6\") pod \"redhat-marketplace-ngw4l\" (UID: \"8fbea3c3-376b-48e1-a863-3eb8478a3974\") " pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:38:49 crc kubenswrapper[4884]: I1202 01:38:49.968273 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.061572 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fbea3c3-376b-48e1-a863-3eb8478a3974-utilities\") pod \"redhat-marketplace-ngw4l\" (UID: \"8fbea3c3-376b-48e1-a863-3eb8478a3974\") " pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.061630 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.061659 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fbea3c3-376b-48e1-a863-3eb8478a3974-catalog-content\") pod \"redhat-marketplace-ngw4l\" (UID: \"8fbea3c3-376b-48e1-a863-3eb8478a3974\") " pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.061703 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f8x6\" (UniqueName: \"kubernetes.io/projected/8fbea3c3-376b-48e1-a863-3eb8478a3974-kube-api-access-2f8x6\") pod \"redhat-marketplace-ngw4l\" (UID: \"8fbea3c3-376b-48e1-a863-3eb8478a3974\") " pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.062244 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fbea3c3-376b-48e1-a863-3eb8478a3974-utilities\") pod \"redhat-marketplace-ngw4l\" (UID: \"8fbea3c3-376b-48e1-a863-3eb8478a3974\") " pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.062654 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fbea3c3-376b-48e1-a863-3eb8478a3974-catalog-content\") pod \"redhat-marketplace-ngw4l\" (UID: \"8fbea3c3-376b-48e1-a863-3eb8478a3974\") " pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.085507 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f8x6\" (UniqueName: \"kubernetes.io/projected/8fbea3c3-376b-48e1-a863-3eb8478a3974-kube-api-access-2f8x6\") pod \"redhat-marketplace-ngw4l\" (UID: \"8fbea3c3-376b-48e1-a863-3eb8478a3974\") " pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.095982 4884 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.096034 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.130984 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-cmk67\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.218191 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7v6v"] Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.234080 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.313806 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:50 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:50 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:50 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.313845 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.377528 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.437962 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ngw4l"] Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.457931 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" event={"ID":"bb7500d3-5d34-451d-9b19-75ae46e5bf97","Type":"ContainerStarted","Data":"b177b4e614db90156046317d2a7f7082107b9ef650ba622f834f0fc71b638901"} Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.457979 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" event={"ID":"bb7500d3-5d34-451d-9b19-75ae46e5bf97","Type":"ContainerStarted","Data":"8accc81519add751e039e356b0413a4bf69c472e37a5df144cae2df890daa8a2"} Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.459410 4884 generic.go:334] "Generic (PLEG): container finished" podID="76949538-95e6-4f55-a9e3-3dcccd40580e" containerID="b1e237eb5e71cd598c294c3337c8468a869a10dbf6222386fc6961e62e16157e" exitCode=0 Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.459453 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vm82" event={"ID":"76949538-95e6-4f55-a9e3-3dcccd40580e","Type":"ContainerDied","Data":"b1e237eb5e71cd598c294c3337c8468a869a10dbf6222386fc6961e62e16157e"} Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.460969 4884 generic.go:334] "Generic (PLEG): container finished" podID="39284c28-156f-445d-9951-9df6851c9487" containerID="1e708c2a239d813d1a696d593e45cb01d59ee99720ccb3a9de898c17cddb5079" exitCode=0 Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.461002 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26msg" event={"ID":"39284c28-156f-445d-9951-9df6851c9487","Type":"ContainerDied","Data":"1e708c2a239d813d1a696d593e45cb01d59ee99720ccb3a9de898c17cddb5079"} Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.462551 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7v6v" event={"ID":"25c18b0c-a129-43c2-8d8e-8381d421510a","Type":"ContainerStarted","Data":"1ffda3e310519bc3f003497cefb591b20a42ab90c9adbd89cb1abf4a3037ac2b"} Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.464642 4884 generic.go:334] "Generic (PLEG): container finished" podID="70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" containerID="a1774190940f7913f4531396bd52b132b0f1f6578b889bfed8aa7b93bc56aa12" exitCode=0 Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.465878 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mzck" event={"ID":"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80","Type":"ContainerDied","Data":"a1774190940f7913f4531396bd52b132b0f1f6578b889bfed8aa7b93bc56aa12"} Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.522331 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jz4cj"] Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.523298 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.524838 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.540344 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jz4cj"] Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.567855 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-utilities\") pod \"redhat-operators-jz4cj\" (UID: \"2ef9b26a-4331-4c67-a38a-921f05f5e8c6\") " pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.567953 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-catalog-content\") pod \"redhat-operators-jz4cj\" (UID: \"2ef9b26a-4331-4c67-a38a-921f05f5e8c6\") " pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.567995 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc8b4\" (UniqueName: \"kubernetes.io/projected/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-kube-api-access-fc8b4\") pod \"redhat-operators-jz4cj\" (UID: \"2ef9b26a-4331-4c67-a38a-921f05f5e8c6\") " pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.604408 4884 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.669555 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.669612 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.669635 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.669667 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-utilities\") pod \"redhat-operators-jz4cj\" (UID: \"2ef9b26a-4331-4c67-a38a-921f05f5e8c6\") " pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.669689 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.669713 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-catalog-content\") pod \"redhat-operators-jz4cj\" (UID: \"2ef9b26a-4331-4c67-a38a-921f05f5e8c6\") " pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.669755 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc8b4\" (UniqueName: \"kubernetes.io/projected/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-kube-api-access-fc8b4\") pod \"redhat-operators-jz4cj\" (UID: \"2ef9b26a-4331-4c67-a38a-921f05f5e8c6\") " pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.670771 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-utilities\") pod \"redhat-operators-jz4cj\" (UID: \"2ef9b26a-4331-4c67-a38a-921f05f5e8c6\") " pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.671967 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-catalog-content\") pod \"redhat-operators-jz4cj\" (UID: \"2ef9b26a-4331-4c67-a38a-921f05f5e8c6\") " pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.675778 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.675830 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.676568 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.684573 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc8b4\" (UniqueName: \"kubernetes.io/projected/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-kube-api-access-fc8b4\") pod \"redhat-operators-jz4cj\" (UID: \"2ef9b26a-4331-4c67-a38a-921f05f5e8c6\") " pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.690309 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.722341 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dxjsm"] Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.723782 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.731260 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dxjsm"] Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.771429 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8518501-8647-42d5-9e20-6d69f40f1e5d-utilities\") pod \"redhat-operators-dxjsm\" (UID: \"e8518501-8647-42d5-9e20-6d69f40f1e5d\") " pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.771506 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8518501-8647-42d5-9e20-6d69f40f1e5d-catalog-content\") pod \"redhat-operators-dxjsm\" (UID: \"e8518501-8647-42d5-9e20-6d69f40f1e5d\") " pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.771551 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ljg9\" (UniqueName: \"kubernetes.io/projected/e8518501-8647-42d5-9e20-6d69f40f1e5d-kube-api-access-9ljg9\") pod \"redhat-operators-dxjsm\" (UID: \"e8518501-8647-42d5-9e20-6d69f40f1e5d\") " pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.831560 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cmk67"] Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.840518 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.843980 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.860007 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.873068 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8518501-8647-42d5-9e20-6d69f40f1e5d-utilities\") pod \"redhat-operators-dxjsm\" (UID: \"e8518501-8647-42d5-9e20-6d69f40f1e5d\") " pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.873157 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8518501-8647-42d5-9e20-6d69f40f1e5d-catalog-content\") pod \"redhat-operators-dxjsm\" (UID: \"e8518501-8647-42d5-9e20-6d69f40f1e5d\") " pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.873219 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ljg9\" (UniqueName: \"kubernetes.io/projected/e8518501-8647-42d5-9e20-6d69f40f1e5d-kube-api-access-9ljg9\") pod \"redhat-operators-dxjsm\" (UID: \"e8518501-8647-42d5-9e20-6d69f40f1e5d\") " pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.874740 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8518501-8647-42d5-9e20-6d69f40f1e5d-utilities\") pod \"redhat-operators-dxjsm\" (UID: \"e8518501-8647-42d5-9e20-6d69f40f1e5d\") " pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.875195 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8518501-8647-42d5-9e20-6d69f40f1e5d-catalog-content\") pod \"redhat-operators-dxjsm\" (UID: \"e8518501-8647-42d5-9e20-6d69f40f1e5d\") " pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.875470 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 02 01:38:50 crc kubenswrapper[4884]: I1202 01:38:50.890395 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ljg9\" (UniqueName: \"kubernetes.io/projected/e8518501-8647-42d5-9e20-6d69f40f1e5d-kube-api-access-9ljg9\") pod \"redhat-operators-dxjsm\" (UID: \"e8518501-8647-42d5-9e20-6d69f40f1e5d\") " pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:38:50 crc kubenswrapper[4884]: W1202 01:38:50.933250 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode17db01a_f964_4380_ac3f_bcad065313e9.slice/crio-442de4d2157002bc2666559a3e5775b7fb0c258d18295c719746e2be8f42687d WatchSource:0}: Error finding container 442de4d2157002bc2666559a3e5775b7fb0c258d18295c719746e2be8f42687d: Status 404 returned error can't find the container with id 442de4d2157002bc2666559a3e5775b7fb0c258d18295c719746e2be8f42687d Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.047163 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.207939 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jz4cj"] Dec 02 01:38:51 crc kubenswrapper[4884]: W1202 01:38:51.237963 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2ef9b26a_4331_4c67_a38a_921f05f5e8c6.slice/crio-4158e42de7e6e5c77c26d34d00d269e648fa03cf0b2357d075442d3823541eec WatchSource:0}: Error finding container 4158e42de7e6e5c77c26d34d00d269e648fa03cf0b2357d075442d3823541eec: Status 404 returned error can't find the container with id 4158e42de7e6e5c77c26d34d00d269e648fa03cf0b2357d075442d3823541eec Dec 02 01:38:51 crc kubenswrapper[4884]: W1202 01:38:51.297662 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-3fa8767ad7c8472794335a507bc1f626f46d5885c6e89ec6092fc2baa91a9697 WatchSource:0}: Error finding container 3fa8767ad7c8472794335a507bc1f626f46d5885c6e89ec6092fc2baa91a9697: Status 404 returned error can't find the container with id 3fa8767ad7c8472794335a507bc1f626f46d5885c6e89ec6092fc2baa91a9697 Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.320819 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:51 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:51 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:51 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.320877 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.476003 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" event={"ID":"e17db01a-f964-4380-ac3f-bcad065313e9","Type":"ContainerStarted","Data":"a86ad0320703025fc535b8991bb4c3d83c21b93817ea1a882d58ca9bfdf53693"} Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.476039 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" event={"ID":"e17db01a-f964-4380-ac3f-bcad065313e9","Type":"ContainerStarted","Data":"442de4d2157002bc2666559a3e5775b7fb0c258d18295c719746e2be8f42687d"} Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.476073 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.479153 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0d33ddd1797f34dec37138dc4de94eb8b3d553cd1ae5d54ed7e26f0330c72166"} Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.479469 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"54ccb8ee7a898a77fd03a624192f53a8c8c6f20f354bc6fb8728af7331d5712c"} Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.480027 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.481437 4884 generic.go:334] "Generic (PLEG): container finished" podID="83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5" containerID="1f29e1274403e52b260ee461e939515a522aeb1b9ea0882d6e1b619c53638712" exitCode=0 Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.481461 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" event={"ID":"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5","Type":"ContainerDied","Data":"1f29e1274403e52b260ee461e939515a522aeb1b9ea0882d6e1b619c53638712"} Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.483411 4884 generic.go:334] "Generic (PLEG): container finished" podID="2ef9b26a-4331-4c67-a38a-921f05f5e8c6" containerID="4887c8c8c8c34c575d931495a2a8cb256102927aae2e3c93010b5af1fd724ef8" exitCode=0 Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.483461 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jz4cj" event={"ID":"2ef9b26a-4331-4c67-a38a-921f05f5e8c6","Type":"ContainerDied","Data":"4887c8c8c8c34c575d931495a2a8cb256102927aae2e3c93010b5af1fd724ef8"} Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.483480 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jz4cj" event={"ID":"2ef9b26a-4331-4c67-a38a-921f05f5e8c6","Type":"ContainerStarted","Data":"4158e42de7e6e5c77c26d34d00d269e648fa03cf0b2357d075442d3823541eec"} Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.489760 4884 generic.go:334] "Generic (PLEG): container finished" podID="25c18b0c-a129-43c2-8d8e-8381d421510a" containerID="7c60316071285a21c46e7bcceff9a79b80407dfe0777a0f26216ab7c8d45eb30" exitCode=0 Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.489860 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7v6v" event={"ID":"25c18b0c-a129-43c2-8d8e-8381d421510a","Type":"ContainerDied","Data":"7c60316071285a21c46e7bcceff9a79b80407dfe0777a0f26216ab7c8d45eb30"} Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.497534 4884 generic.go:334] "Generic (PLEG): container finished" podID="8fbea3c3-376b-48e1-a863-3eb8478a3974" containerID="7ba26a340b820cfd80806b461af03b0a6e45b64331229bf0768fc000517dade9" exitCode=0 Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.498510 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngw4l" event={"ID":"8fbea3c3-376b-48e1-a863-3eb8478a3974","Type":"ContainerDied","Data":"7ba26a340b820cfd80806b461af03b0a6e45b64331229bf0768fc000517dade9"} Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.498574 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngw4l" event={"ID":"8fbea3c3-376b-48e1-a863-3eb8478a3974","Type":"ContainerStarted","Data":"37211b7113998cfa41b3ba13f1be1e38990656494ce85fa503c53a7ca34e2168"} Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.505567 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" podStartSLOduration=129.505543218 podStartE2EDuration="2m9.505543218s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:51.498470476 +0000 UTC m=+148.174307400" watchObservedRunningTime="2025-12-02 01:38:51.505543218 +0000 UTC m=+148.181380122" Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.508623 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"3b69feb7d6486ab9f8aaae6129eb002bb4434f57fc184a9a75c2f7128f967993"} Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.508666 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"3fa8767ad7c8472794335a507bc1f626f46d5885c6e89ec6092fc2baa91a9697"} Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.612618 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dxjsm"] Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.634677 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-6gvqj" podStartSLOduration=12.634650787 podStartE2EDuration="12.634650787s" podCreationTimestamp="2025-12-02 01:38:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:51.630030505 +0000 UTC m=+148.305867419" watchObservedRunningTime="2025-12-02 01:38:51.634650787 +0000 UTC m=+148.310487671" Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.635230 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 02 01:38:51 crc kubenswrapper[4884]: W1202 01:38:51.638670 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8518501_8647_42d5_9e20_6d69f40f1e5d.slice/crio-fef356f69b35d70ea56d70e664d5d8917fd389ce284e561b9676b096c400668b WatchSource:0}: Error finding container fef356f69b35d70ea56d70e664d5d8917fd389ce284e561b9676b096c400668b: Status 404 returned error can't find the container with id fef356f69b35d70ea56d70e664d5d8917fd389ce284e561b9676b096c400668b Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.911580 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.912204 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.914444 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.914640 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.916498 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.917780 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.917823 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.919957 4884 patch_prober.go:28] interesting pod/console-f9d7485db-mh86q container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.920006 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mh86q" podUID="d3232fbf-e182-4e1b-a27d-ae2cf61b40b7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.991983 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f31ffc21-cc7a-4588-8207-fe49efb85141-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f31ffc21-cc7a-4588-8207-fe49efb85141\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 01:38:51 crc kubenswrapper[4884]: I1202 01:38:51.993255 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f31ffc21-cc7a-4588-8207-fe49efb85141-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f31ffc21-cc7a-4588-8207-fe49efb85141\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.095669 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f31ffc21-cc7a-4588-8207-fe49efb85141-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f31ffc21-cc7a-4588-8207-fe49efb85141\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.095818 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f31ffc21-cc7a-4588-8207-fe49efb85141-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f31ffc21-cc7a-4588-8207-fe49efb85141\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.096137 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f31ffc21-cc7a-4588-8207-fe49efb85141-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"f31ffc21-cc7a-4588-8207-fe49efb85141\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.120515 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f31ffc21-cc7a-4588-8207-fe49efb85141-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"f31ffc21-cc7a-4588-8207-fe49efb85141\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.158202 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.158278 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.165031 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.226780 4884 patch_prober.go:28] interesting pod/downloads-7954f5f757-pmdpb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.226864 4884 patch_prober.go:28] interesting pod/downloads-7954f5f757-pmdpb container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.226921 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-pmdpb" podUID="22ccfdc9-0d00-475d-ac54-d084beaca9dd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.226895 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-pmdpb" podUID="22ccfdc9-0d00-475d-ac54-d084beaca9dd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.237976 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.319688 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.322989 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:52 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:52 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:52 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.323060 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.528235 4884 generic.go:334] "Generic (PLEG): container finished" podID="e8518501-8647-42d5-9e20-6d69f40f1e5d" containerID="55daefc06977efef0a26f4409684163a3883a073dde2199a040514f272ebc88c" exitCode=0 Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.528604 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxjsm" event={"ID":"e8518501-8647-42d5-9e20-6d69f40f1e5d","Type":"ContainerDied","Data":"55daefc06977efef0a26f4409684163a3883a073dde2199a040514f272ebc88c"} Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.528641 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxjsm" event={"ID":"e8518501-8647-42d5-9e20-6d69f40f1e5d","Type":"ContainerStarted","Data":"fef356f69b35d70ea56d70e664d5d8917fd389ce284e561b9676b096c400668b"} Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.533832 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"9955285690fcdcc70d439feb39d28443369173a0b6477af272ce19d0d8fe5af4"} Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.533918 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2ffb318596e10b9d5334af38b45fe0e6b489a35c52d4be576d229bb7ef48d620"} Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.537859 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-sfn66" Dec 02 01:38:52 crc kubenswrapper[4884]: I1202 01:38:52.555764 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.071206 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.216423 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-secret-volume\") pod \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\" (UID: \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\") " Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.216526 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-config-volume\") pod \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\" (UID: \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\") " Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.216595 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77cbq\" (UniqueName: \"kubernetes.io/projected/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-kube-api-access-77cbq\") pod \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\" (UID: \"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5\") " Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.217575 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-config-volume" (OuterVolumeSpecName: "config-volume") pod "83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5" (UID: "83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.237941 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-kube-api-access-77cbq" (OuterVolumeSpecName: "kube-api-access-77cbq") pod "83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5" (UID: "83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5"). InnerVolumeSpecName "kube-api-access-77cbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.237998 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5" (UID: "83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.320591 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:53 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:53 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:53 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.320687 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.321991 4884 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.322035 4884 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.322052 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77cbq\" (UniqueName: \"kubernetes.io/projected/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5-kube-api-access-77cbq\") on node \"crc\" DevicePath \"\"" Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.552474 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f31ffc21-cc7a-4588-8207-fe49efb85141","Type":"ContainerStarted","Data":"7869e53711e242f9f5c425eb363d9e0eabe8c39084c34d7056a5fe3bfa31720d"} Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.552549 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f31ffc21-cc7a-4588-8207-fe49efb85141","Type":"ContainerStarted","Data":"dc82abf33b76e5df86adc73d3e213e6f24b7b21511c2e88eb605d9c2ce4b4c06"} Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.559326 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.559489 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4" event={"ID":"83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5","Type":"ContainerDied","Data":"95d7f03a4fda579f3c2b0c64be1a8624875c185a22ce386a96d3773a04c83fbd"} Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.559527 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95d7f03a4fda579f3c2b0c64be1a8624875c185a22ce386a96d3773a04c83fbd" Dec 02 01:38:53 crc kubenswrapper[4884]: I1202 01:38:53.623470 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.623132751 podStartE2EDuration="2.623132751s" podCreationTimestamp="2025-12-02 01:38:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:53.567545289 +0000 UTC m=+150.243382173" watchObservedRunningTime="2025-12-02 01:38:53.623132751 +0000 UTC m=+150.298969645" Dec 02 01:38:54 crc kubenswrapper[4884]: I1202 01:38:54.316151 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:54 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:54 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:54 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:54 crc kubenswrapper[4884]: I1202 01:38:54.316531 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:38:54 crc kubenswrapper[4884]: I1202 01:38:54.571453 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f31ffc21-cc7a-4588-8207-fe49efb85141","Type":"ContainerDied","Data":"7869e53711e242f9f5c425eb363d9e0eabe8c39084c34d7056a5fe3bfa31720d"} Dec 02 01:38:54 crc kubenswrapper[4884]: I1202 01:38:54.571434 4884 generic.go:334] "Generic (PLEG): container finished" podID="f31ffc21-cc7a-4588-8207-fe49efb85141" containerID="7869e53711e242f9f5c425eb363d9e0eabe8c39084c34d7056a5fe3bfa31720d" exitCode=0 Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.311325 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:55 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:55 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:55 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.311402 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.720050 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 02 01:38:55 crc kubenswrapper[4884]: E1202 01:38:55.720273 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5" containerName="collect-profiles" Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.720284 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5" containerName="collect-profiles" Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.720375 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5" containerName="collect-profiles" Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.720807 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.729315 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.729366 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.731767 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.861921 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9bd5d35-c035-4c74-bda9-340102cae87f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e9bd5d35-c035-4c74-bda9-340102cae87f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.862035 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e9bd5d35-c035-4c74-bda9-340102cae87f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e9bd5d35-c035-4c74-bda9-340102cae87f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.926127 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.964577 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e9bd5d35-c035-4c74-bda9-340102cae87f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e9bd5d35-c035-4c74-bda9-340102cae87f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.964669 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e9bd5d35-c035-4c74-bda9-340102cae87f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e9bd5d35-c035-4c74-bda9-340102cae87f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.964672 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9bd5d35-c035-4c74-bda9-340102cae87f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e9bd5d35-c035-4c74-bda9-340102cae87f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 01:38:55 crc kubenswrapper[4884]: I1202 01:38:55.981830 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9bd5d35-c035-4c74-bda9-340102cae87f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e9bd5d35-c035-4c74-bda9-340102cae87f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 01:38:56 crc kubenswrapper[4884]: I1202 01:38:56.054120 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 01:38:56 crc kubenswrapper[4884]: I1202 01:38:56.065491 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f31ffc21-cc7a-4588-8207-fe49efb85141-kubelet-dir\") pod \"f31ffc21-cc7a-4588-8207-fe49efb85141\" (UID: \"f31ffc21-cc7a-4588-8207-fe49efb85141\") " Dec 02 01:38:56 crc kubenswrapper[4884]: I1202 01:38:56.065612 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f31ffc21-cc7a-4588-8207-fe49efb85141-kube-api-access\") pod \"f31ffc21-cc7a-4588-8207-fe49efb85141\" (UID: \"f31ffc21-cc7a-4588-8207-fe49efb85141\") " Dec 02 01:38:56 crc kubenswrapper[4884]: I1202 01:38:56.065924 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f31ffc21-cc7a-4588-8207-fe49efb85141-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f31ffc21-cc7a-4588-8207-fe49efb85141" (UID: "f31ffc21-cc7a-4588-8207-fe49efb85141"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:38:56 crc kubenswrapper[4884]: I1202 01:38:56.068592 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f31ffc21-cc7a-4588-8207-fe49efb85141-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f31ffc21-cc7a-4588-8207-fe49efb85141" (UID: "f31ffc21-cc7a-4588-8207-fe49efb85141"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:38:56 crc kubenswrapper[4884]: I1202 01:38:56.167363 4884 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f31ffc21-cc7a-4588-8207-fe49efb85141-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 02 01:38:56 crc kubenswrapper[4884]: I1202 01:38:56.167400 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f31ffc21-cc7a-4588-8207-fe49efb85141-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 01:38:56 crc kubenswrapper[4884]: I1202 01:38:56.312444 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:56 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:56 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:56 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:56 crc kubenswrapper[4884]: I1202 01:38:56.312499 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:38:56 crc kubenswrapper[4884]: I1202 01:38:56.437056 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 02 01:38:56 crc kubenswrapper[4884]: I1202 01:38:56.595653 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e9bd5d35-c035-4c74-bda9-340102cae87f","Type":"ContainerStarted","Data":"b8205d4b7f56bf30b73fad9d5541ab5f5560f0f92f00548c3450642a7e9e4ecd"} Dec 02 01:38:56 crc kubenswrapper[4884]: I1202 01:38:56.597660 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"f31ffc21-cc7a-4588-8207-fe49efb85141","Type":"ContainerDied","Data":"dc82abf33b76e5df86adc73d3e213e6f24b7b21511c2e88eb605d9c2ce4b4c06"} Dec 02 01:38:56 crc kubenswrapper[4884]: I1202 01:38:56.597688 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc82abf33b76e5df86adc73d3e213e6f24b7b21511c2e88eb605d9c2ce4b4c06" Dec 02 01:38:56 crc kubenswrapper[4884]: I1202 01:38:56.597809 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 02 01:38:57 crc kubenswrapper[4884]: I1202 01:38:57.312284 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:57 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:57 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:57 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:57 crc kubenswrapper[4884]: I1202 01:38:57.312395 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:38:57 crc kubenswrapper[4884]: I1202 01:38:57.626963 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e9bd5d35-c035-4c74-bda9-340102cae87f","Type":"ContainerStarted","Data":"3a80a2f84e87ce300358cfd27375c1b53960ec6b7285c46f1e486d7226c123a2"} Dec 02 01:38:57 crc kubenswrapper[4884]: I1202 01:38:57.644646 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.644626939 podStartE2EDuration="2.644626939s" podCreationTimestamp="2025-12-02 01:38:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:38:57.641011641 +0000 UTC m=+154.316848525" watchObservedRunningTime="2025-12-02 01:38:57.644626939 +0000 UTC m=+154.320463823" Dec 02 01:38:57 crc kubenswrapper[4884]: I1202 01:38:57.798375 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-74x48" Dec 02 01:38:58 crc kubenswrapper[4884]: I1202 01:38:58.312757 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:58 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:58 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:58 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:58 crc kubenswrapper[4884]: I1202 01:38:58.312820 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:38:58 crc kubenswrapper[4884]: I1202 01:38:58.640389 4884 generic.go:334] "Generic (PLEG): container finished" podID="e9bd5d35-c035-4c74-bda9-340102cae87f" containerID="3a80a2f84e87ce300358cfd27375c1b53960ec6b7285c46f1e486d7226c123a2" exitCode=0 Dec 02 01:38:58 crc kubenswrapper[4884]: I1202 01:38:58.640436 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e9bd5d35-c035-4c74-bda9-340102cae87f","Type":"ContainerDied","Data":"3a80a2f84e87ce300358cfd27375c1b53960ec6b7285c46f1e486d7226c123a2"} Dec 02 01:38:59 crc kubenswrapper[4884]: I1202 01:38:59.314853 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:38:59 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:38:59 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:38:59 crc kubenswrapper[4884]: healthz check failed Dec 02 01:38:59 crc kubenswrapper[4884]: I1202 01:38:59.314922 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:39:00 crc kubenswrapper[4884]: I1202 01:39:00.312452 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:39:00 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:39:00 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:39:00 crc kubenswrapper[4884]: healthz check failed Dec 02 01:39:00 crc kubenswrapper[4884]: I1202 01:39:00.312645 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:39:01 crc kubenswrapper[4884]: I1202 01:39:01.311208 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:39:01 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:39:01 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:39:01 crc kubenswrapper[4884]: healthz check failed Dec 02 01:39:01 crc kubenswrapper[4884]: I1202 01:39:01.311261 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:39:01 crc kubenswrapper[4884]: I1202 01:39:01.918705 4884 patch_prober.go:28] interesting pod/console-f9d7485db-mh86q container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 02 01:39:01 crc kubenswrapper[4884]: I1202 01:39:01.918785 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mh86q" podUID="d3232fbf-e182-4e1b-a27d-ae2cf61b40b7" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 02 01:39:02 crc kubenswrapper[4884]: I1202 01:39:02.243253 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-pmdpb" Dec 02 01:39:02 crc kubenswrapper[4884]: I1202 01:39:02.316955 4884 patch_prober.go:28] interesting pod/router-default-5444994796-664p9 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 02 01:39:02 crc kubenswrapper[4884]: [-]has-synced failed: reason withheld Dec 02 01:39:02 crc kubenswrapper[4884]: [+]process-running ok Dec 02 01:39:02 crc kubenswrapper[4884]: healthz check failed Dec 02 01:39:02 crc kubenswrapper[4884]: I1202 01:39:02.317041 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-664p9" podUID="52dd5b9f-cc6f-4898-a5dd-a235270c39e0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 02 01:39:04 crc kubenswrapper[4884]: I1202 01:39:04.285093 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:39:04 crc kubenswrapper[4884]: I1202 01:39:04.295799 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-664p9" Dec 02 01:39:04 crc kubenswrapper[4884]: I1202 01:39:04.801438 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs\") pod \"network-metrics-daemon-whh6t\" (UID: \"7a6a67e3-3b98-45df-9573-6efee410a90c\") " pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:39:04 crc kubenswrapper[4884]: I1202 01:39:04.810085 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a6a67e3-3b98-45df-9573-6efee410a90c-metrics-certs\") pod \"network-metrics-daemon-whh6t\" (UID: \"7a6a67e3-3b98-45df-9573-6efee410a90c\") " pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:39:05 crc kubenswrapper[4884]: I1202 01:39:05.037477 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-whh6t" Dec 02 01:39:09 crc kubenswrapper[4884]: I1202 01:39:09.177182 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 01:39:09 crc kubenswrapper[4884]: I1202 01:39:09.265491 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9bd5d35-c035-4c74-bda9-340102cae87f-kube-api-access\") pod \"e9bd5d35-c035-4c74-bda9-340102cae87f\" (UID: \"e9bd5d35-c035-4c74-bda9-340102cae87f\") " Dec 02 01:39:09 crc kubenswrapper[4884]: I1202 01:39:09.265645 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e9bd5d35-c035-4c74-bda9-340102cae87f-kubelet-dir\") pod \"e9bd5d35-c035-4c74-bda9-340102cae87f\" (UID: \"e9bd5d35-c035-4c74-bda9-340102cae87f\") " Dec 02 01:39:09 crc kubenswrapper[4884]: I1202 01:39:09.266046 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9bd5d35-c035-4c74-bda9-340102cae87f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e9bd5d35-c035-4c74-bda9-340102cae87f" (UID: "e9bd5d35-c035-4c74-bda9-340102cae87f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:39:09 crc kubenswrapper[4884]: I1202 01:39:09.272934 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9bd5d35-c035-4c74-bda9-340102cae87f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e9bd5d35-c035-4c74-bda9-340102cae87f" (UID: "e9bd5d35-c035-4c74-bda9-340102cae87f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:39:09 crc kubenswrapper[4884]: I1202 01:39:09.368065 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e9bd5d35-c035-4c74-bda9-340102cae87f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 01:39:09 crc kubenswrapper[4884]: I1202 01:39:09.368125 4884 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e9bd5d35-c035-4c74-bda9-340102cae87f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 02 01:39:09 crc kubenswrapper[4884]: I1202 01:39:09.711728 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e9bd5d35-c035-4c74-bda9-340102cae87f","Type":"ContainerDied","Data":"b8205d4b7f56bf30b73fad9d5541ab5f5560f0f92f00548c3450642a7e9e4ecd"} Dec 02 01:39:09 crc kubenswrapper[4884]: I1202 01:39:09.711829 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 02 01:39:09 crc kubenswrapper[4884]: I1202 01:39:09.711865 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8205d4b7f56bf30b73fad9d5541ab5f5560f0f92f00548c3450642a7e9e4ecd" Dec 02 01:39:10 crc kubenswrapper[4884]: I1202 01:39:10.386313 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:39:11 crc kubenswrapper[4884]: I1202 01:39:11.961406 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:39:11 crc kubenswrapper[4884]: I1202 01:39:11.964758 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:39:16 crc kubenswrapper[4884]: I1202 01:39:16.971830 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:39:16 crc kubenswrapper[4884]: I1202 01:39:16.972140 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:39:22 crc kubenswrapper[4884]: I1202 01:39:22.651143 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wkddb" Dec 02 01:39:26 crc kubenswrapper[4884]: E1202 01:39:26.786142 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 02 01:39:26 crc kubenswrapper[4884]: E1202 01:39:26.786669 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m2dmz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-2mzck_openshift-marketplace(70d5f628-082b-4f7b-8cc9-c4a08e5d8e80): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 01:39:26 crc kubenswrapper[4884]: E1202 01:39:26.787907 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-2mzck" podUID="70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.126457 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 02 01:39:28 crc kubenswrapper[4884]: E1202 01:39:28.126689 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9bd5d35-c035-4c74-bda9-340102cae87f" containerName="pruner" Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.126700 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9bd5d35-c035-4c74-bda9-340102cae87f" containerName="pruner" Dec 02 01:39:28 crc kubenswrapper[4884]: E1202 01:39:28.126708 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f31ffc21-cc7a-4588-8207-fe49efb85141" containerName="pruner" Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.126716 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f31ffc21-cc7a-4588-8207-fe49efb85141" containerName="pruner" Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.126827 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9bd5d35-c035-4c74-bda9-340102cae87f" containerName="pruner" Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.126844 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="f31ffc21-cc7a-4588-8207-fe49efb85141" containerName="pruner" Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.127138 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.127212 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.130539 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.130763 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.242596 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2b22bdb-961e-4920-841a-2d5f4e352c2a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a2b22bdb-961e-4920-841a-2d5f4e352c2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.242657 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2b22bdb-961e-4920-841a-2d5f4e352c2a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a2b22bdb-961e-4920-841a-2d5f4e352c2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.343806 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2b22bdb-961e-4920-841a-2d5f4e352c2a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a2b22bdb-961e-4920-841a-2d5f4e352c2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.343891 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2b22bdb-961e-4920-841a-2d5f4e352c2a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a2b22bdb-961e-4920-841a-2d5f4e352c2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.343957 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2b22bdb-961e-4920-841a-2d5f4e352c2a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a2b22bdb-961e-4920-841a-2d5f4e352c2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.474971 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2b22bdb-961e-4920-841a-2d5f4e352c2a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a2b22bdb-961e-4920-841a-2d5f4e352c2a\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 01:39:28 crc kubenswrapper[4884]: I1202 01:39:28.757815 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 01:39:30 crc kubenswrapper[4884]: E1202 01:39:30.060679 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-2mzck" podUID="70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" Dec 02 01:39:30 crc kubenswrapper[4884]: E1202 01:39:30.252459 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 02 01:39:30 crc kubenswrapper[4884]: E1202 01:39:30.252727 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9ljg9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-dxjsm_openshift-marketplace(e8518501-8647-42d5-9e20-6d69f40f1e5d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 01:39:30 crc kubenswrapper[4884]: E1202 01:39:30.253969 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-dxjsm" podUID="e8518501-8647-42d5-9e20-6d69f40f1e5d" Dec 02 01:39:30 crc kubenswrapper[4884]: E1202 01:39:30.516986 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 02 01:39:30 crc kubenswrapper[4884]: E1202 01:39:30.517276 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fc8b4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-jz4cj_openshift-marketplace(2ef9b26a-4331-4c67-a38a-921f05f5e8c6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 01:39:30 crc kubenswrapper[4884]: E1202 01:39:30.518943 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-jz4cj" podUID="2ef9b26a-4331-4c67-a38a-921f05f5e8c6" Dec 02 01:39:31 crc kubenswrapper[4884]: I1202 01:39:31.457037 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 02 01:39:31 crc kubenswrapper[4884]: E1202 01:39:31.490094 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-jz4cj" podUID="2ef9b26a-4331-4c67-a38a-921f05f5e8c6" Dec 02 01:39:31 crc kubenswrapper[4884]: E1202 01:39:31.491230 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-dxjsm" podUID="e8518501-8647-42d5-9e20-6d69f40f1e5d" Dec 02 01:39:31 crc kubenswrapper[4884]: E1202 01:39:31.580566 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 02 01:39:31 crc kubenswrapper[4884]: E1202 01:39:31.581016 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2f8x6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-ngw4l_openshift-marketplace(8fbea3c3-376b-48e1-a863-3eb8478a3974): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 01:39:31 crc kubenswrapper[4884]: E1202 01:39:31.582301 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-ngw4l" podUID="8fbea3c3-376b-48e1-a863-3eb8478a3974" Dec 02 01:39:31 crc kubenswrapper[4884]: E1202 01:39:31.627472 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 02 01:39:31 crc kubenswrapper[4884]: E1202 01:39:31.627628 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-829qm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-26msg_openshift-marketplace(39284c28-156f-445d-9951-9df6851c9487): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 01:39:31 crc kubenswrapper[4884]: E1202 01:39:31.628867 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-26msg" podUID="39284c28-156f-445d-9951-9df6851c9487" Dec 02 01:39:31 crc kubenswrapper[4884]: I1202 01:39:31.737146 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-whh6t"] Dec 02 01:39:32 crc kubenswrapper[4884]: E1202 01:39:32.995225 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-ngw4l" podUID="8fbea3c3-376b-48e1-a863-3eb8478a3974" Dec 02 01:39:32 crc kubenswrapper[4884]: E1202 01:39:32.995225 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-26msg" podUID="39284c28-156f-445d-9951-9df6851c9487" Dec 02 01:39:33 crc kubenswrapper[4884]: W1202 01:39:33.006424 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a6a67e3_3b98_45df_9573_6efee410a90c.slice/crio-e44902b3af592e2e75de3ee529334c21236aa2e4bcfa113f9a64a9dd30f1bd8f WatchSource:0}: Error finding container e44902b3af592e2e75de3ee529334c21236aa2e4bcfa113f9a64a9dd30f1bd8f: Status 404 returned error can't find the container with id e44902b3af592e2e75de3ee529334c21236aa2e4bcfa113f9a64a9dd30f1bd8f Dec 02 01:39:33 crc kubenswrapper[4884]: E1202 01:39:33.061727 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 02 01:39:33 crc kubenswrapper[4884]: E1202 01:39:33.062243 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pg8v8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-tqtn7_openshift-marketplace(525da77f-9387-441b-8ded-8f7b1cc41df8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 01:39:33 crc kubenswrapper[4884]: E1202 01:39:33.063618 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-tqtn7" podUID="525da77f-9387-441b-8ded-8f7b1cc41df8" Dec 02 01:39:33 crc kubenswrapper[4884]: E1202 01:39:33.077626 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 02 01:39:33 crc kubenswrapper[4884]: E1202 01:39:33.080044 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vxzgn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-6vm82_openshift-marketplace(76949538-95e6-4f55-a9e3-3dcccd40580e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 01:39:33 crc kubenswrapper[4884]: E1202 01:39:33.082831 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-6vm82" podUID="76949538-95e6-4f55-a9e3-3dcccd40580e" Dec 02 01:39:33 crc kubenswrapper[4884]: E1202 01:39:33.116418 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 02 01:39:33 crc kubenswrapper[4884]: E1202 01:39:33.116586 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7tx4l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-k7v6v_openshift-marketplace(25c18b0c-a129-43c2-8d8e-8381d421510a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 02 01:39:33 crc kubenswrapper[4884]: E1202 01:39:33.117787 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-k7v6v" podUID="25c18b0c-a129-43c2-8d8e-8381d421510a" Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.199047 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.508026 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.508922 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.520523 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.641181 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ca937f60-dd07-4e98-a9aa-1ab893593165-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ca937f60-dd07-4e98-a9aa-1ab893593165\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.641265 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ca937f60-dd07-4e98-a9aa-1ab893593165-kube-api-access\") pod \"installer-9-crc\" (UID: \"ca937f60-dd07-4e98-a9aa-1ab893593165\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.641464 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ca937f60-dd07-4e98-a9aa-1ab893593165-var-lock\") pod \"installer-9-crc\" (UID: \"ca937f60-dd07-4e98-a9aa-1ab893593165\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.742634 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ca937f60-dd07-4e98-a9aa-1ab893593165-var-lock\") pod \"installer-9-crc\" (UID: \"ca937f60-dd07-4e98-a9aa-1ab893593165\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.742692 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ca937f60-dd07-4e98-a9aa-1ab893593165-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ca937f60-dd07-4e98-a9aa-1ab893593165\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.742765 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ca937f60-dd07-4e98-a9aa-1ab893593165-kube-api-access\") pod \"installer-9-crc\" (UID: \"ca937f60-dd07-4e98-a9aa-1ab893593165\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.742778 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ca937f60-dd07-4e98-a9aa-1ab893593165-kubelet-dir\") pod \"installer-9-crc\" (UID: \"ca937f60-dd07-4e98-a9aa-1ab893593165\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.742779 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ca937f60-dd07-4e98-a9aa-1ab893593165-var-lock\") pod \"installer-9-crc\" (UID: \"ca937f60-dd07-4e98-a9aa-1ab893593165\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.771988 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ca937f60-dd07-4e98-a9aa-1ab893593165-kube-api-access\") pod \"installer-9-crc\" (UID: \"ca937f60-dd07-4e98-a9aa-1ab893593165\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.839948 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.861878 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-whh6t" event={"ID":"7a6a67e3-3b98-45df-9573-6efee410a90c","Type":"ContainerStarted","Data":"3ef193e5a795a7aeb00f59b9aebad5372a6fcf80fb13b333c94cb4788581e26a"} Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.861931 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-whh6t" event={"ID":"7a6a67e3-3b98-45df-9573-6efee410a90c","Type":"ContainerStarted","Data":"e812dc2f8b29daf68a88f1d1d4d2533a33f7fbf36ab8f3ec1fdf0acff77e75ff"} Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.861945 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-whh6t" event={"ID":"7a6a67e3-3b98-45df-9573-6efee410a90c","Type":"ContainerStarted","Data":"e44902b3af592e2e75de3ee529334c21236aa2e4bcfa113f9a64a9dd30f1bd8f"} Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.863655 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a2b22bdb-961e-4920-841a-2d5f4e352c2a","Type":"ContainerStarted","Data":"ffb9fd166f1f961d5beef66167a0e00d3b445cf6037a97b0ece5758469690233"} Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.863713 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a2b22bdb-961e-4920-841a-2d5f4e352c2a","Type":"ContainerStarted","Data":"0b4bcae2c7aa9a891984ab8d95aa7f4365666f9b491f8b353951e0528fd04d92"} Dec 02 01:39:33 crc kubenswrapper[4884]: E1202 01:39:33.869647 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-6vm82" podUID="76949538-95e6-4f55-a9e3-3dcccd40580e" Dec 02 01:39:33 crc kubenswrapper[4884]: E1202 01:39:33.869916 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-k7v6v" podUID="25c18b0c-a129-43c2-8d8e-8381d421510a" Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.885568 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-whh6t" podStartSLOduration=171.885548404 podStartE2EDuration="2m51.885548404s" podCreationTimestamp="2025-12-02 01:36:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:39:33.884608692 +0000 UTC m=+190.560445616" watchObservedRunningTime="2025-12-02 01:39:33.885548404 +0000 UTC m=+190.561385308" Dec 02 01:39:33 crc kubenswrapper[4884]: I1202 01:39:33.909807 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=5.909785234 podStartE2EDuration="5.909785234s" podCreationTimestamp="2025-12-02 01:39:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:39:33.900073698 +0000 UTC m=+190.575910642" watchObservedRunningTime="2025-12-02 01:39:33.909785234 +0000 UTC m=+190.585622128" Dec 02 01:39:34 crc kubenswrapper[4884]: I1202 01:39:34.115441 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 02 01:39:34 crc kubenswrapper[4884]: I1202 01:39:34.871566 4884 generic.go:334] "Generic (PLEG): container finished" podID="a2b22bdb-961e-4920-841a-2d5f4e352c2a" containerID="ffb9fd166f1f961d5beef66167a0e00d3b445cf6037a97b0ece5758469690233" exitCode=0 Dec 02 01:39:34 crc kubenswrapper[4884]: I1202 01:39:34.871660 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a2b22bdb-961e-4920-841a-2d5f4e352c2a","Type":"ContainerDied","Data":"ffb9fd166f1f961d5beef66167a0e00d3b445cf6037a97b0ece5758469690233"} Dec 02 01:39:34 crc kubenswrapper[4884]: I1202 01:39:34.873794 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ca937f60-dd07-4e98-a9aa-1ab893593165","Type":"ContainerStarted","Data":"05dc4b2aa5878107884bd06f3617013fc5917ed427e2ab6dec4aaecceb1a7dc9"} Dec 02 01:39:34 crc kubenswrapper[4884]: I1202 01:39:34.873824 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ca937f60-dd07-4e98-a9aa-1ab893593165","Type":"ContainerStarted","Data":"914d31058d7b22435d96a6027e1e4dd84c2353d177e03a1e6ae9cb8eba2e995e"} Dec 02 01:39:34 crc kubenswrapper[4884]: I1202 01:39:34.899879 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.899858489 podStartE2EDuration="1.899858489s" podCreationTimestamp="2025-12-02 01:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:39:34.898034685 +0000 UTC m=+191.573871619" watchObservedRunningTime="2025-12-02 01:39:34.899858489 +0000 UTC m=+191.575695373" Dec 02 01:39:36 crc kubenswrapper[4884]: I1202 01:39:36.178664 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 01:39:36 crc kubenswrapper[4884]: I1202 01:39:36.274148 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2b22bdb-961e-4920-841a-2d5f4e352c2a-kube-api-access\") pod \"a2b22bdb-961e-4920-841a-2d5f4e352c2a\" (UID: \"a2b22bdb-961e-4920-841a-2d5f4e352c2a\") " Dec 02 01:39:36 crc kubenswrapper[4884]: I1202 01:39:36.274202 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2b22bdb-961e-4920-841a-2d5f4e352c2a-kubelet-dir\") pod \"a2b22bdb-961e-4920-841a-2d5f4e352c2a\" (UID: \"a2b22bdb-961e-4920-841a-2d5f4e352c2a\") " Dec 02 01:39:36 crc kubenswrapper[4884]: I1202 01:39:36.274392 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a2b22bdb-961e-4920-841a-2d5f4e352c2a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a2b22bdb-961e-4920-841a-2d5f4e352c2a" (UID: "a2b22bdb-961e-4920-841a-2d5f4e352c2a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:39:36 crc kubenswrapper[4884]: I1202 01:39:36.274780 4884 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a2b22bdb-961e-4920-841a-2d5f4e352c2a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 02 01:39:36 crc kubenswrapper[4884]: I1202 01:39:36.279151 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2b22bdb-961e-4920-841a-2d5f4e352c2a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a2b22bdb-961e-4920-841a-2d5f4e352c2a" (UID: "a2b22bdb-961e-4920-841a-2d5f4e352c2a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:39:36 crc kubenswrapper[4884]: I1202 01:39:36.376477 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a2b22bdb-961e-4920-841a-2d5f4e352c2a-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 01:39:36 crc kubenswrapper[4884]: I1202 01:39:36.893912 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a2b22bdb-961e-4920-841a-2d5f4e352c2a","Type":"ContainerDied","Data":"0b4bcae2c7aa9a891984ab8d95aa7f4365666f9b491f8b353951e0528fd04d92"} Dec 02 01:39:36 crc kubenswrapper[4884]: I1202 01:39:36.893949 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b4bcae2c7aa9a891984ab8d95aa7f4365666f9b491f8b353951e0528fd04d92" Dec 02 01:39:36 crc kubenswrapper[4884]: I1202 01:39:36.894020 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 02 01:39:44 crc kubenswrapper[4884]: I1202 01:39:44.938669 4884 generic.go:334] "Generic (PLEG): container finished" podID="70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" containerID="1e671dbaf37ae0d9e526c73efb5e65d452d25ae29aa8028e8617bae03cbd397e" exitCode=0 Dec 02 01:39:44 crc kubenswrapper[4884]: I1202 01:39:44.938773 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mzck" event={"ID":"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80","Type":"ContainerDied","Data":"1e671dbaf37ae0d9e526c73efb5e65d452d25ae29aa8028e8617bae03cbd397e"} Dec 02 01:39:45 crc kubenswrapper[4884]: I1202 01:39:45.945706 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mzck" event={"ID":"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80","Type":"ContainerStarted","Data":"f8c4d24df9a9b1d334b6eb38f35e5137358b21c757f512c41c4b1e3528992b88"} Dec 02 01:39:45 crc kubenswrapper[4884]: I1202 01:39:45.947493 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxjsm" event={"ID":"e8518501-8647-42d5-9e20-6d69f40f1e5d","Type":"ContainerStarted","Data":"be83f2fc00055884e59a2a5e4a087616b352c64b2c242d1c577fdce01ae42f00"} Dec 02 01:39:45 crc kubenswrapper[4884]: I1202 01:39:45.963782 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2mzck" podStartSLOduration=4.204057845 podStartE2EDuration="58.963765317s" podCreationTimestamp="2025-12-02 01:38:47 +0000 UTC" firstStartedPulling="2025-12-02 01:38:50.605583075 +0000 UTC m=+147.281419959" lastFinishedPulling="2025-12-02 01:39:45.365290547 +0000 UTC m=+202.041127431" observedRunningTime="2025-12-02 01:39:45.961375669 +0000 UTC m=+202.637212573" watchObservedRunningTime="2025-12-02 01:39:45.963765317 +0000 UTC m=+202.639602191" Dec 02 01:39:46 crc kubenswrapper[4884]: I1202 01:39:46.953121 4884 generic.go:334] "Generic (PLEG): container finished" podID="e8518501-8647-42d5-9e20-6d69f40f1e5d" containerID="be83f2fc00055884e59a2a5e4a087616b352c64b2c242d1c577fdce01ae42f00" exitCode=0 Dec 02 01:39:46 crc kubenswrapper[4884]: I1202 01:39:46.953185 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxjsm" event={"ID":"e8518501-8647-42d5-9e20-6d69f40f1e5d","Type":"ContainerDied","Data":"be83f2fc00055884e59a2a5e4a087616b352c64b2c242d1c577fdce01ae42f00"} Dec 02 01:39:46 crc kubenswrapper[4884]: I1202 01:39:46.955104 4884 generic.go:334] "Generic (PLEG): container finished" podID="39284c28-156f-445d-9951-9df6851c9487" containerID="7c6647edbe9cdb96cde74f5ce26065780368988f008809eb07061ac7def93df7" exitCode=0 Dec 02 01:39:46 crc kubenswrapper[4884]: I1202 01:39:46.955155 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26msg" event={"ID":"39284c28-156f-445d-9951-9df6851c9487","Type":"ContainerDied","Data":"7c6647edbe9cdb96cde74f5ce26065780368988f008809eb07061ac7def93df7"} Dec 02 01:39:46 crc kubenswrapper[4884]: I1202 01:39:46.956829 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jz4cj" event={"ID":"2ef9b26a-4331-4c67-a38a-921f05f5e8c6","Type":"ContainerStarted","Data":"29b31ec9e2fcb4ef073b210ded678e61ed3288333b6a4f8a6728257960bb3781"} Dec 02 01:39:46 crc kubenswrapper[4884]: I1202 01:39:46.959947 4884 generic.go:334] "Generic (PLEG): container finished" podID="8fbea3c3-376b-48e1-a863-3eb8478a3974" containerID="37c050f7aaafb8a7cb65c42b8717edba7340bfa0e9cee9a110c888fd69e20753" exitCode=0 Dec 02 01:39:46 crc kubenswrapper[4884]: I1202 01:39:46.959975 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngw4l" event={"ID":"8fbea3c3-376b-48e1-a863-3eb8478a3974","Type":"ContainerDied","Data":"37c050f7aaafb8a7cb65c42b8717edba7340bfa0e9cee9a110c888fd69e20753"} Dec 02 01:39:46 crc kubenswrapper[4884]: I1202 01:39:46.971357 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:39:46 crc kubenswrapper[4884]: I1202 01:39:46.971420 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:39:46 crc kubenswrapper[4884]: I1202 01:39:46.971466 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:39:46 crc kubenswrapper[4884]: I1202 01:39:46.972127 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 01:39:46 crc kubenswrapper[4884]: I1202 01:39:46.972253 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f" gracePeriod=600 Dec 02 01:39:47 crc kubenswrapper[4884]: I1202 01:39:47.868484 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:39:47 crc kubenswrapper[4884]: I1202 01:39:47.869082 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:39:47 crc kubenswrapper[4884]: I1202 01:39:47.929277 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:39:47 crc kubenswrapper[4884]: I1202 01:39:47.966553 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26msg" event={"ID":"39284c28-156f-445d-9951-9df6851c9487","Type":"ContainerStarted","Data":"a3e520cb40452822aa9bcac8f85fe8fdb5690c30ea86a0d7be194559f5f40716"} Dec 02 01:39:47 crc kubenswrapper[4884]: I1202 01:39:47.968056 4884 generic.go:334] "Generic (PLEG): container finished" podID="2ef9b26a-4331-4c67-a38a-921f05f5e8c6" containerID="29b31ec9e2fcb4ef073b210ded678e61ed3288333b6a4f8a6728257960bb3781" exitCode=0 Dec 02 01:39:47 crc kubenswrapper[4884]: I1202 01:39:47.968116 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jz4cj" event={"ID":"2ef9b26a-4331-4c67-a38a-921f05f5e8c6","Type":"ContainerDied","Data":"29b31ec9e2fcb4ef073b210ded678e61ed3288333b6a4f8a6728257960bb3781"} Dec 02 01:39:47 crc kubenswrapper[4884]: I1202 01:39:47.970682 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngw4l" event={"ID":"8fbea3c3-376b-48e1-a863-3eb8478a3974","Type":"ContainerStarted","Data":"8ffc2e3cbdd292272c7569c70ce985a0034e913795e4fc282f7d8aae43cb40eb"} Dec 02 01:39:47 crc kubenswrapper[4884]: I1202 01:39:47.972982 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f" exitCode=0 Dec 02 01:39:47 crc kubenswrapper[4884]: I1202 01:39:47.973038 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f"} Dec 02 01:39:47 crc kubenswrapper[4884]: I1202 01:39:47.973056 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"e037a7c9f1b0c94849181495610a0b1c6c18fa025d1d93c587fd598c55910e9f"} Dec 02 01:39:47 crc kubenswrapper[4884]: I1202 01:39:47.975669 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxjsm" event={"ID":"e8518501-8647-42d5-9e20-6d69f40f1e5d","Type":"ContainerStarted","Data":"08f5b1beafb3ee5cabf176fae6e8816debca17d7824936022670596a3abe86e5"} Dec 02 01:39:47 crc kubenswrapper[4884]: I1202 01:39:47.986221 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-26msg" podStartSLOduration=4.257691389 podStartE2EDuration="1m0.986204775s" podCreationTimestamp="2025-12-02 01:38:47 +0000 UTC" firstStartedPulling="2025-12-02 01:38:50.605093533 +0000 UTC m=+147.280930417" lastFinishedPulling="2025-12-02 01:39:47.333606909 +0000 UTC m=+204.009443803" observedRunningTime="2025-12-02 01:39:47.98395203 +0000 UTC m=+204.659788914" watchObservedRunningTime="2025-12-02 01:39:47.986204775 +0000 UTC m=+204.662041669" Dec 02 01:39:48 crc kubenswrapper[4884]: I1202 01:39:48.037518 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ngw4l" podStartSLOduration=3.071764811 podStartE2EDuration="59.03749941s" podCreationTimestamp="2025-12-02 01:38:49 +0000 UTC" firstStartedPulling="2025-12-02 01:38:51.505976788 +0000 UTC m=+148.181813692" lastFinishedPulling="2025-12-02 01:39:47.471711397 +0000 UTC m=+204.147548291" observedRunningTime="2025-12-02 01:39:48.03624904 +0000 UTC m=+204.712085924" watchObservedRunningTime="2025-12-02 01:39:48.03749941 +0000 UTC m=+204.713336294" Dec 02 01:39:48 crc kubenswrapper[4884]: I1202 01:39:48.053302 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dxjsm" podStartSLOduration=3.084146429 podStartE2EDuration="58.053285634s" podCreationTimestamp="2025-12-02 01:38:50 +0000 UTC" firstStartedPulling="2025-12-02 01:38:52.531981467 +0000 UTC m=+149.207818351" lastFinishedPulling="2025-12-02 01:39:47.501120672 +0000 UTC m=+204.176957556" observedRunningTime="2025-12-02 01:39:48.050799594 +0000 UTC m=+204.726636508" watchObservedRunningTime="2025-12-02 01:39:48.053285634 +0000 UTC m=+204.729122518" Dec 02 01:39:48 crc kubenswrapper[4884]: I1202 01:39:48.436010 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:39:48 crc kubenswrapper[4884]: I1202 01:39:48.436060 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:39:48 crc kubenswrapper[4884]: I1202 01:39:48.982117 4884 generic.go:334] "Generic (PLEG): container finished" podID="76949538-95e6-4f55-a9e3-3dcccd40580e" containerID="1b4ae6f344dd37b71a8ac7e8775d232a89bc87ae68f0b7d6e611c5e755faa0c0" exitCode=0 Dec 02 01:39:48 crc kubenswrapper[4884]: I1202 01:39:48.982218 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vm82" event={"ID":"76949538-95e6-4f55-a9e3-3dcccd40580e","Type":"ContainerDied","Data":"1b4ae6f344dd37b71a8ac7e8775d232a89bc87ae68f0b7d6e611c5e755faa0c0"} Dec 02 01:39:48 crc kubenswrapper[4884]: I1202 01:39:48.984451 4884 generic.go:334] "Generic (PLEG): container finished" podID="525da77f-9387-441b-8ded-8f7b1cc41df8" containerID="1b962ae3c89230dbba89742da2888e20ca6e6d0f08a8a4b868ce6b167108e7c5" exitCode=0 Dec 02 01:39:48 crc kubenswrapper[4884]: I1202 01:39:48.984506 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqtn7" event={"ID":"525da77f-9387-441b-8ded-8f7b1cc41df8","Type":"ContainerDied","Data":"1b962ae3c89230dbba89742da2888e20ca6e6d0f08a8a4b868ce6b167108e7c5"} Dec 02 01:39:48 crc kubenswrapper[4884]: I1202 01:39:48.986464 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jz4cj" event={"ID":"2ef9b26a-4331-4c67-a38a-921f05f5e8c6","Type":"ContainerStarted","Data":"ffd58676dac6f722261f8badb316a8139566b56398b9ae2b5b76e50be89f2b0e"} Dec 02 01:39:49 crc kubenswrapper[4884]: I1202 01:39:49.046366 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jz4cj" podStartSLOduration=2.111349896 podStartE2EDuration="59.046348887s" podCreationTimestamp="2025-12-02 01:38:50 +0000 UTC" firstStartedPulling="2025-12-02 01:38:51.489654102 +0000 UTC m=+148.165490976" lastFinishedPulling="2025-12-02 01:39:48.424653083 +0000 UTC m=+205.100489967" observedRunningTime="2025-12-02 01:39:49.04523505 +0000 UTC m=+205.721071934" watchObservedRunningTime="2025-12-02 01:39:49.046348887 +0000 UTC m=+205.722185791" Dec 02 01:39:49 crc kubenswrapper[4884]: I1202 01:39:49.484498 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-26msg" podUID="39284c28-156f-445d-9951-9df6851c9487" containerName="registry-server" probeResult="failure" output=< Dec 02 01:39:49 crc kubenswrapper[4884]: timeout: failed to connect service ":50051" within 1s Dec 02 01:39:49 crc kubenswrapper[4884]: > Dec 02 01:39:50 crc kubenswrapper[4884]: I1202 01:39:50.234438 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:39:50 crc kubenswrapper[4884]: I1202 01:39:50.234836 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:39:50 crc kubenswrapper[4884]: I1202 01:39:50.281355 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:39:50 crc kubenswrapper[4884]: I1202 01:39:50.844111 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:39:50 crc kubenswrapper[4884]: I1202 01:39:50.844180 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:39:50 crc kubenswrapper[4884]: I1202 01:39:50.998268 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vm82" event={"ID":"76949538-95e6-4f55-a9e3-3dcccd40580e","Type":"ContainerStarted","Data":"f570cb2595d6f895190507961e109ada053c3148e6f7213cb9cc7543a3fb422a"} Dec 02 01:39:51 crc kubenswrapper[4884]: I1202 01:39:51.000980 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqtn7" event={"ID":"525da77f-9387-441b-8ded-8f7b1cc41df8","Type":"ContainerStarted","Data":"c35594d89bfd93397dda6bc0fb6e4af987d34797d082e57b9f310ebfd560baeb"} Dec 02 01:39:51 crc kubenswrapper[4884]: I1202 01:39:51.018627 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6vm82" podStartSLOduration=4.965785393 podStartE2EDuration="1m4.018609854s" podCreationTimestamp="2025-12-02 01:38:47 +0000 UTC" firstStartedPulling="2025-12-02 01:38:50.605326589 +0000 UTC m=+147.281163473" lastFinishedPulling="2025-12-02 01:39:49.65815105 +0000 UTC m=+206.333987934" observedRunningTime="2025-12-02 01:39:51.017783514 +0000 UTC m=+207.693620428" watchObservedRunningTime="2025-12-02 01:39:51.018609854 +0000 UTC m=+207.694446748" Dec 02 01:39:51 crc kubenswrapper[4884]: I1202 01:39:51.040009 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tqtn7" podStartSLOduration=4.944396873 podStartE2EDuration="1m4.039992244s" podCreationTimestamp="2025-12-02 01:38:47 +0000 UTC" firstStartedPulling="2025-12-02 01:38:50.604043678 +0000 UTC m=+147.279880552" lastFinishedPulling="2025-12-02 01:39:49.699639039 +0000 UTC m=+206.375475923" observedRunningTime="2025-12-02 01:39:51.036431287 +0000 UTC m=+207.712268171" watchObservedRunningTime="2025-12-02 01:39:51.039992244 +0000 UTC m=+207.715829128" Dec 02 01:39:51 crc kubenswrapper[4884]: I1202 01:39:51.048792 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:39:51 crc kubenswrapper[4884]: I1202 01:39:51.048823 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:39:51 crc kubenswrapper[4884]: I1202 01:39:51.890159 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jz4cj" podUID="2ef9b26a-4331-4c67-a38a-921f05f5e8c6" containerName="registry-server" probeResult="failure" output=< Dec 02 01:39:51 crc kubenswrapper[4884]: timeout: failed to connect service ":50051" within 1s Dec 02 01:39:51 crc kubenswrapper[4884]: > Dec 02 01:39:52 crc kubenswrapper[4884]: I1202 01:39:52.009712 4884 generic.go:334] "Generic (PLEG): container finished" podID="25c18b0c-a129-43c2-8d8e-8381d421510a" containerID="92221f554ae5054dc1ef6472f2f4cad275510a4e5186db81750a20598d4f2ea6" exitCode=0 Dec 02 01:39:52 crc kubenswrapper[4884]: I1202 01:39:52.009781 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7v6v" event={"ID":"25c18b0c-a129-43c2-8d8e-8381d421510a","Type":"ContainerDied","Data":"92221f554ae5054dc1ef6472f2f4cad275510a4e5186db81750a20598d4f2ea6"} Dec 02 01:39:52 crc kubenswrapper[4884]: I1202 01:39:52.087124 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dxjsm" podUID="e8518501-8647-42d5-9e20-6d69f40f1e5d" containerName="registry-server" probeResult="failure" output=< Dec 02 01:39:52 crc kubenswrapper[4884]: timeout: failed to connect service ":50051" within 1s Dec 02 01:39:52 crc kubenswrapper[4884]: > Dec 02 01:39:54 crc kubenswrapper[4884]: I1202 01:39:54.021138 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7v6v" event={"ID":"25c18b0c-a129-43c2-8d8e-8381d421510a","Type":"ContainerStarted","Data":"76ff6d9afa54717a7d32fc7480cf0547d36d368de1a04634a492db3b9075f3f0"} Dec 02 01:39:54 crc kubenswrapper[4884]: I1202 01:39:54.042472 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k7v6v" podStartSLOduration=3.381353499 podStartE2EDuration="1m5.042453446s" podCreationTimestamp="2025-12-02 01:38:49 +0000 UTC" firstStartedPulling="2025-12-02 01:38:51.493060844 +0000 UTC m=+148.168897728" lastFinishedPulling="2025-12-02 01:39:53.154160771 +0000 UTC m=+209.829997675" observedRunningTime="2025-12-02 01:39:54.039940625 +0000 UTC m=+210.715777519" watchObservedRunningTime="2025-12-02 01:39:54.042453446 +0000 UTC m=+210.718290340" Dec 02 01:39:57 crc kubenswrapper[4884]: I1202 01:39:57.655307 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:39:57 crc kubenswrapper[4884]: I1202 01:39:57.655965 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:39:57 crc kubenswrapper[4884]: I1202 01:39:57.731203 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:39:57 crc kubenswrapper[4884]: I1202 01:39:57.908762 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:39:58 crc kubenswrapper[4884]: I1202 01:39:58.111666 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:39:58 crc kubenswrapper[4884]: I1202 01:39:58.134269 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:39:58 crc kubenswrapper[4884]: I1202 01:39:58.134333 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:39:58 crc kubenswrapper[4884]: I1202 01:39:58.188145 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:39:58 crc kubenswrapper[4884]: I1202 01:39:58.489591 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:39:58 crc kubenswrapper[4884]: I1202 01:39:58.530551 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:39:59 crc kubenswrapper[4884]: I1202 01:39:59.095003 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:39:59 crc kubenswrapper[4884]: I1202 01:39:59.447189 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-26msg"] Dec 02 01:39:59 crc kubenswrapper[4884]: I1202 01:39:59.951872 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:39:59 crc kubenswrapper[4884]: I1202 01:39:59.951936 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:40:00 crc kubenswrapper[4884]: I1202 01:40:00.021255 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:40:00 crc kubenswrapper[4884]: I1202 01:40:00.059254 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-26msg" podUID="39284c28-156f-445d-9951-9df6851c9487" containerName="registry-server" containerID="cri-o://a3e520cb40452822aa9bcac8f85fe8fdb5690c30ea86a0d7be194559f5f40716" gracePeriod=2 Dec 02 01:40:00 crc kubenswrapper[4884]: I1202 01:40:00.114169 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:40:00 crc kubenswrapper[4884]: I1202 01:40:00.287706 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:40:00 crc kubenswrapper[4884]: I1202 01:40:00.445429 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6vm82"] Dec 02 01:40:00 crc kubenswrapper[4884]: I1202 01:40:00.914138 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:40:00 crc kubenswrapper[4884]: I1202 01:40:00.968336 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:40:01 crc kubenswrapper[4884]: I1202 01:40:01.064643 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6vm82" podUID="76949538-95e6-4f55-a9e3-3dcccd40580e" containerName="registry-server" containerID="cri-o://f570cb2595d6f895190507961e109ada053c3148e6f7213cb9cc7543a3fb422a" gracePeriod=2 Dec 02 01:40:01 crc kubenswrapper[4884]: I1202 01:40:01.105312 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:40:01 crc kubenswrapper[4884]: I1202 01:40:01.152014 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:40:02 crc kubenswrapper[4884]: I1202 01:40:02.847341 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ngw4l"] Dec 02 01:40:02 crc kubenswrapper[4884]: I1202 01:40:02.849172 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ngw4l" podUID="8fbea3c3-376b-48e1-a863-3eb8478a3974" containerName="registry-server" containerID="cri-o://8ffc2e3cbdd292272c7569c70ce985a0034e913795e4fc282f7d8aae43cb40eb" gracePeriod=2 Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.037201 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wpxzm"] Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.116358 4884 generic.go:334] "Generic (PLEG): container finished" podID="76949538-95e6-4f55-a9e3-3dcccd40580e" containerID="f570cb2595d6f895190507961e109ada053c3148e6f7213cb9cc7543a3fb422a" exitCode=0 Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.116437 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vm82" event={"ID":"76949538-95e6-4f55-a9e3-3dcccd40580e","Type":"ContainerDied","Data":"f570cb2595d6f895190507961e109ada053c3148e6f7213cb9cc7543a3fb422a"} Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.126039 4884 generic.go:334] "Generic (PLEG): container finished" podID="39284c28-156f-445d-9951-9df6851c9487" containerID="a3e520cb40452822aa9bcac8f85fe8fdb5690c30ea86a0d7be194559f5f40716" exitCode=0 Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.126079 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26msg" event={"ID":"39284c28-156f-445d-9951-9df6851c9487","Type":"ContainerDied","Data":"a3e520cb40452822aa9bcac8f85fe8fdb5690c30ea86a0d7be194559f5f40716"} Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.127537 4884 generic.go:334] "Generic (PLEG): container finished" podID="8fbea3c3-376b-48e1-a863-3eb8478a3974" containerID="8ffc2e3cbdd292272c7569c70ce985a0034e913795e4fc282f7d8aae43cb40eb" exitCode=0 Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.127558 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngw4l" event={"ID":"8fbea3c3-376b-48e1-a863-3eb8478a3974","Type":"ContainerDied","Data":"8ffc2e3cbdd292272c7569c70ce985a0034e913795e4fc282f7d8aae43cb40eb"} Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.381458 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.382856 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.504202 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fbea3c3-376b-48e1-a863-3eb8478a3974-catalog-content\") pod \"8fbea3c3-376b-48e1-a863-3eb8478a3974\" (UID: \"8fbea3c3-376b-48e1-a863-3eb8478a3974\") " Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.504288 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f8x6\" (UniqueName: \"kubernetes.io/projected/8fbea3c3-376b-48e1-a863-3eb8478a3974-kube-api-access-2f8x6\") pod \"8fbea3c3-376b-48e1-a863-3eb8478a3974\" (UID: \"8fbea3c3-376b-48e1-a863-3eb8478a3974\") " Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.505098 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-829qm\" (UniqueName: \"kubernetes.io/projected/39284c28-156f-445d-9951-9df6851c9487-kube-api-access-829qm\") pod \"39284c28-156f-445d-9951-9df6851c9487\" (UID: \"39284c28-156f-445d-9951-9df6851c9487\") " Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.505248 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39284c28-156f-445d-9951-9df6851c9487-utilities\") pod \"39284c28-156f-445d-9951-9df6851c9487\" (UID: \"39284c28-156f-445d-9951-9df6851c9487\") " Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.505333 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39284c28-156f-445d-9951-9df6851c9487-catalog-content\") pod \"39284c28-156f-445d-9951-9df6851c9487\" (UID: \"39284c28-156f-445d-9951-9df6851c9487\") " Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.505387 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fbea3c3-376b-48e1-a863-3eb8478a3974-utilities\") pod \"8fbea3c3-376b-48e1-a863-3eb8478a3974\" (UID: \"8fbea3c3-376b-48e1-a863-3eb8478a3974\") " Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.505913 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39284c28-156f-445d-9951-9df6851c9487-utilities" (OuterVolumeSpecName: "utilities") pod "39284c28-156f-445d-9951-9df6851c9487" (UID: "39284c28-156f-445d-9951-9df6851c9487"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.506031 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39284c28-156f-445d-9951-9df6851c9487-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.507184 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fbea3c3-376b-48e1-a863-3eb8478a3974-utilities" (OuterVolumeSpecName: "utilities") pod "8fbea3c3-376b-48e1-a863-3eb8478a3974" (UID: "8fbea3c3-376b-48e1-a863-3eb8478a3974"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.510094 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fbea3c3-376b-48e1-a863-3eb8478a3974-kube-api-access-2f8x6" (OuterVolumeSpecName: "kube-api-access-2f8x6") pod "8fbea3c3-376b-48e1-a863-3eb8478a3974" (UID: "8fbea3c3-376b-48e1-a863-3eb8478a3974"). InnerVolumeSpecName "kube-api-access-2f8x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.511696 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39284c28-156f-445d-9951-9df6851c9487-kube-api-access-829qm" (OuterVolumeSpecName: "kube-api-access-829qm") pod "39284c28-156f-445d-9951-9df6851c9487" (UID: "39284c28-156f-445d-9951-9df6851c9487"). InnerVolumeSpecName "kube-api-access-829qm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.541323 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fbea3c3-376b-48e1-a863-3eb8478a3974-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8fbea3c3-376b-48e1-a863-3eb8478a3974" (UID: "8fbea3c3-376b-48e1-a863-3eb8478a3974"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.583072 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39284c28-156f-445d-9951-9df6851c9487-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "39284c28-156f-445d-9951-9df6851c9487" (UID: "39284c28-156f-445d-9951-9df6851c9487"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.608185 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39284c28-156f-445d-9951-9df6851c9487-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.608507 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fbea3c3-376b-48e1-a863-3eb8478a3974-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.608621 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fbea3c3-376b-48e1-a863-3eb8478a3974-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.608814 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2f8x6\" (UniqueName: \"kubernetes.io/projected/8fbea3c3-376b-48e1-a863-3eb8478a3974-kube-api-access-2f8x6\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:04 crc kubenswrapper[4884]: I1202 01:40:04.608949 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-829qm\" (UniqueName: \"kubernetes.io/projected/39284c28-156f-445d-9951-9df6851c9487-kube-api-access-829qm\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.140290 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6vm82" event={"ID":"76949538-95e6-4f55-a9e3-3dcccd40580e","Type":"ContainerDied","Data":"47e8c5a549e12fa67e0ed20e228d780aa8d2b6d720c700dbfe5d60e8c49478f4"} Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.140980 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47e8c5a549e12fa67e0ed20e228d780aa8d2b6d720c700dbfe5d60e8c49478f4" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.143637 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-26msg" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.143564 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-26msg" event={"ID":"39284c28-156f-445d-9951-9df6851c9487","Type":"ContainerDied","Data":"73bf0058ddf2441a7b2d9a2890ec92b4c6023d9db35c2b2977efb7e0d8301f8a"} Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.144526 4884 scope.go:117] "RemoveContainer" containerID="a3e520cb40452822aa9bcac8f85fe8fdb5690c30ea86a0d7be194559f5f40716" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.147739 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ngw4l" event={"ID":"8fbea3c3-376b-48e1-a863-3eb8478a3974","Type":"ContainerDied","Data":"37211b7113998cfa41b3ba13f1be1e38990656494ce85fa503c53a7ca34e2168"} Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.148034 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ngw4l" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.196293 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.200485 4884 scope.go:117] "RemoveContainer" containerID="7c6647edbe9cdb96cde74f5ce26065780368988f008809eb07061ac7def93df7" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.209162 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-26msg"] Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.213794 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-26msg"] Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.220872 4884 scope.go:117] "RemoveContainer" containerID="1e708c2a239d813d1a696d593e45cb01d59ee99720ccb3a9de898c17cddb5079" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.242105 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ngw4l"] Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.246438 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ngw4l"] Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.248824 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dxjsm"] Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.249041 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dxjsm" podUID="e8518501-8647-42d5-9e20-6d69f40f1e5d" containerName="registry-server" containerID="cri-o://08f5b1beafb3ee5cabf176fae6e8816debca17d7824936022670596a3abe86e5" gracePeriod=2 Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.250288 4884 scope.go:117] "RemoveContainer" containerID="8ffc2e3cbdd292272c7569c70ce985a0034e913795e4fc282f7d8aae43cb40eb" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.273109 4884 scope.go:117] "RemoveContainer" containerID="37c050f7aaafb8a7cb65c42b8717edba7340bfa0e9cee9a110c888fd69e20753" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.287100 4884 scope.go:117] "RemoveContainer" containerID="7ba26a340b820cfd80806b461af03b0a6e45b64331229bf0768fc000517dade9" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.317287 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76949538-95e6-4f55-a9e3-3dcccd40580e-catalog-content\") pod \"76949538-95e6-4f55-a9e3-3dcccd40580e\" (UID: \"76949538-95e6-4f55-a9e3-3dcccd40580e\") " Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.317350 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxzgn\" (UniqueName: \"kubernetes.io/projected/76949538-95e6-4f55-a9e3-3dcccd40580e-kube-api-access-vxzgn\") pod \"76949538-95e6-4f55-a9e3-3dcccd40580e\" (UID: \"76949538-95e6-4f55-a9e3-3dcccd40580e\") " Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.317392 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76949538-95e6-4f55-a9e3-3dcccd40580e-utilities\") pod \"76949538-95e6-4f55-a9e3-3dcccd40580e\" (UID: \"76949538-95e6-4f55-a9e3-3dcccd40580e\") " Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.318840 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76949538-95e6-4f55-a9e3-3dcccd40580e-utilities" (OuterVolumeSpecName: "utilities") pod "76949538-95e6-4f55-a9e3-3dcccd40580e" (UID: "76949538-95e6-4f55-a9e3-3dcccd40580e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.327432 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76949538-95e6-4f55-a9e3-3dcccd40580e-kube-api-access-vxzgn" (OuterVolumeSpecName: "kube-api-access-vxzgn") pod "76949538-95e6-4f55-a9e3-3dcccd40580e" (UID: "76949538-95e6-4f55-a9e3-3dcccd40580e"). InnerVolumeSpecName "kube-api-access-vxzgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.369652 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76949538-95e6-4f55-a9e3-3dcccd40580e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76949538-95e6-4f55-a9e3-3dcccd40580e" (UID: "76949538-95e6-4f55-a9e3-3dcccd40580e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.420119 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76949538-95e6-4f55-a9e3-3dcccd40580e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.420168 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxzgn\" (UniqueName: \"kubernetes.io/projected/76949538-95e6-4f55-a9e3-3dcccd40580e-kube-api-access-vxzgn\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.420189 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76949538-95e6-4f55-a9e3-3dcccd40580e-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.623055 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39284c28-156f-445d-9951-9df6851c9487" path="/var/lib/kubelet/pods/39284c28-156f-445d-9951-9df6851c9487/volumes" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.624073 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fbea3c3-376b-48e1-a863-3eb8478a3974" path="/var/lib/kubelet/pods/8fbea3c3-376b-48e1-a863-3eb8478a3974/volumes" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.707981 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.824782 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8518501-8647-42d5-9e20-6d69f40f1e5d-catalog-content\") pod \"e8518501-8647-42d5-9e20-6d69f40f1e5d\" (UID: \"e8518501-8647-42d5-9e20-6d69f40f1e5d\") " Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.824872 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ljg9\" (UniqueName: \"kubernetes.io/projected/e8518501-8647-42d5-9e20-6d69f40f1e5d-kube-api-access-9ljg9\") pod \"e8518501-8647-42d5-9e20-6d69f40f1e5d\" (UID: \"e8518501-8647-42d5-9e20-6d69f40f1e5d\") " Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.824913 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8518501-8647-42d5-9e20-6d69f40f1e5d-utilities\") pod \"e8518501-8647-42d5-9e20-6d69f40f1e5d\" (UID: \"e8518501-8647-42d5-9e20-6d69f40f1e5d\") " Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.827046 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8518501-8647-42d5-9e20-6d69f40f1e5d-utilities" (OuterVolumeSpecName: "utilities") pod "e8518501-8647-42d5-9e20-6d69f40f1e5d" (UID: "e8518501-8647-42d5-9e20-6d69f40f1e5d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.830492 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8518501-8647-42d5-9e20-6d69f40f1e5d-kube-api-access-9ljg9" (OuterVolumeSpecName: "kube-api-access-9ljg9") pod "e8518501-8647-42d5-9e20-6d69f40f1e5d" (UID: "e8518501-8647-42d5-9e20-6d69f40f1e5d"). InnerVolumeSpecName "kube-api-access-9ljg9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.927580 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ljg9\" (UniqueName: \"kubernetes.io/projected/e8518501-8647-42d5-9e20-6d69f40f1e5d-kube-api-access-9ljg9\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:05 crc kubenswrapper[4884]: I1202 01:40:05.927681 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8518501-8647-42d5-9e20-6d69f40f1e5d-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.008187 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8518501-8647-42d5-9e20-6d69f40f1e5d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8518501-8647-42d5-9e20-6d69f40f1e5d" (UID: "e8518501-8647-42d5-9e20-6d69f40f1e5d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.029335 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8518501-8647-42d5-9e20-6d69f40f1e5d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.159330 4884 generic.go:334] "Generic (PLEG): container finished" podID="e8518501-8647-42d5-9e20-6d69f40f1e5d" containerID="08f5b1beafb3ee5cabf176fae6e8816debca17d7824936022670596a3abe86e5" exitCode=0 Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.159372 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxjsm" event={"ID":"e8518501-8647-42d5-9e20-6d69f40f1e5d","Type":"ContainerDied","Data":"08f5b1beafb3ee5cabf176fae6e8816debca17d7824936022670596a3abe86e5"} Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.159447 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6vm82" Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.159446 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxjsm" event={"ID":"e8518501-8647-42d5-9e20-6d69f40f1e5d","Type":"ContainerDied","Data":"fef356f69b35d70ea56d70e664d5d8917fd389ce284e561b9676b096c400668b"} Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.159481 4884 scope.go:117] "RemoveContainer" containerID="08f5b1beafb3ee5cabf176fae6e8816debca17d7824936022670596a3abe86e5" Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.159477 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dxjsm" Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.180890 4884 scope.go:117] "RemoveContainer" containerID="be83f2fc00055884e59a2a5e4a087616b352c64b2c242d1c577fdce01ae42f00" Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.190772 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6vm82"] Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.195972 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6vm82"] Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.197844 4884 scope.go:117] "RemoveContainer" containerID="55daefc06977efef0a26f4409684163a3883a073dde2199a040514f272ebc88c" Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.236891 4884 scope.go:117] "RemoveContainer" containerID="08f5b1beafb3ee5cabf176fae6e8816debca17d7824936022670596a3abe86e5" Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.237839 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dxjsm"] Dec 02 01:40:06 crc kubenswrapper[4884]: E1202 01:40:06.239265 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08f5b1beafb3ee5cabf176fae6e8816debca17d7824936022670596a3abe86e5\": container with ID starting with 08f5b1beafb3ee5cabf176fae6e8816debca17d7824936022670596a3abe86e5 not found: ID does not exist" containerID="08f5b1beafb3ee5cabf176fae6e8816debca17d7824936022670596a3abe86e5" Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.239385 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08f5b1beafb3ee5cabf176fae6e8816debca17d7824936022670596a3abe86e5"} err="failed to get container status \"08f5b1beafb3ee5cabf176fae6e8816debca17d7824936022670596a3abe86e5\": rpc error: code = NotFound desc = could not find container \"08f5b1beafb3ee5cabf176fae6e8816debca17d7824936022670596a3abe86e5\": container with ID starting with 08f5b1beafb3ee5cabf176fae6e8816debca17d7824936022670596a3abe86e5 not found: ID does not exist" Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.239482 4884 scope.go:117] "RemoveContainer" containerID="be83f2fc00055884e59a2a5e4a087616b352c64b2c242d1c577fdce01ae42f00" Dec 02 01:40:06 crc kubenswrapper[4884]: E1202 01:40:06.240104 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be83f2fc00055884e59a2a5e4a087616b352c64b2c242d1c577fdce01ae42f00\": container with ID starting with be83f2fc00055884e59a2a5e4a087616b352c64b2c242d1c577fdce01ae42f00 not found: ID does not exist" containerID="be83f2fc00055884e59a2a5e4a087616b352c64b2c242d1c577fdce01ae42f00" Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.240266 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be83f2fc00055884e59a2a5e4a087616b352c64b2c242d1c577fdce01ae42f00"} err="failed to get container status \"be83f2fc00055884e59a2a5e4a087616b352c64b2c242d1c577fdce01ae42f00\": rpc error: code = NotFound desc = could not find container \"be83f2fc00055884e59a2a5e4a087616b352c64b2c242d1c577fdce01ae42f00\": container with ID starting with be83f2fc00055884e59a2a5e4a087616b352c64b2c242d1c577fdce01ae42f00 not found: ID does not exist" Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.240372 4884 scope.go:117] "RemoveContainer" containerID="55daefc06977efef0a26f4409684163a3883a073dde2199a040514f272ebc88c" Dec 02 01:40:06 crc kubenswrapper[4884]: E1202 01:40:06.240897 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55daefc06977efef0a26f4409684163a3883a073dde2199a040514f272ebc88c\": container with ID starting with 55daefc06977efef0a26f4409684163a3883a073dde2199a040514f272ebc88c not found: ID does not exist" containerID="55daefc06977efef0a26f4409684163a3883a073dde2199a040514f272ebc88c" Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.240991 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55daefc06977efef0a26f4409684163a3883a073dde2199a040514f272ebc88c"} err="failed to get container status \"55daefc06977efef0a26f4409684163a3883a073dde2199a040514f272ebc88c\": rpc error: code = NotFound desc = could not find container \"55daefc06977efef0a26f4409684163a3883a073dde2199a040514f272ebc88c\": container with ID starting with 55daefc06977efef0a26f4409684163a3883a073dde2199a040514f272ebc88c not found: ID does not exist" Dec 02 01:40:06 crc kubenswrapper[4884]: I1202 01:40:06.242538 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dxjsm"] Dec 02 01:40:07 crc kubenswrapper[4884]: I1202 01:40:07.627612 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76949538-95e6-4f55-a9e3-3dcccd40580e" path="/var/lib/kubelet/pods/76949538-95e6-4f55-a9e3-3dcccd40580e/volumes" Dec 02 01:40:07 crc kubenswrapper[4884]: I1202 01:40:07.629185 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8518501-8647-42d5-9e20-6d69f40f1e5d" path="/var/lib/kubelet/pods/e8518501-8647-42d5-9e20-6d69f40f1e5d/volumes" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.187393 4884 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.188085 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fbea3c3-376b-48e1-a863-3eb8478a3974" containerName="extract-utilities" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188109 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fbea3c3-376b-48e1-a863-3eb8478a3974" containerName="extract-utilities" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.188129 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8518501-8647-42d5-9e20-6d69f40f1e5d" containerName="registry-server" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188142 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8518501-8647-42d5-9e20-6d69f40f1e5d" containerName="registry-server" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.188160 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b22bdb-961e-4920-841a-2d5f4e352c2a" containerName="pruner" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188174 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b22bdb-961e-4920-841a-2d5f4e352c2a" containerName="pruner" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.188189 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76949538-95e6-4f55-a9e3-3dcccd40580e" containerName="extract-content" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188201 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="76949538-95e6-4f55-a9e3-3dcccd40580e" containerName="extract-content" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.188218 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8518501-8647-42d5-9e20-6d69f40f1e5d" containerName="extract-utilities" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188232 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8518501-8647-42d5-9e20-6d69f40f1e5d" containerName="extract-utilities" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.188256 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39284c28-156f-445d-9951-9df6851c9487" containerName="registry-server" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188268 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="39284c28-156f-445d-9951-9df6851c9487" containerName="registry-server" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.188287 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39284c28-156f-445d-9951-9df6851c9487" containerName="extract-utilities" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188299 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="39284c28-156f-445d-9951-9df6851c9487" containerName="extract-utilities" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.188318 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fbea3c3-376b-48e1-a863-3eb8478a3974" containerName="extract-content" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188332 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fbea3c3-376b-48e1-a863-3eb8478a3974" containerName="extract-content" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.188347 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76949538-95e6-4f55-a9e3-3dcccd40580e" containerName="extract-utilities" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188358 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="76949538-95e6-4f55-a9e3-3dcccd40580e" containerName="extract-utilities" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.188377 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8518501-8647-42d5-9e20-6d69f40f1e5d" containerName="extract-content" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188390 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8518501-8647-42d5-9e20-6d69f40f1e5d" containerName="extract-content" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.188410 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76949538-95e6-4f55-a9e3-3dcccd40580e" containerName="registry-server" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188422 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="76949538-95e6-4f55-a9e3-3dcccd40580e" containerName="registry-server" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.188442 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fbea3c3-376b-48e1-a863-3eb8478a3974" containerName="registry-server" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188453 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fbea3c3-376b-48e1-a863-3eb8478a3974" containerName="registry-server" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.188474 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39284c28-156f-445d-9951-9df6851c9487" containerName="extract-content" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188487 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="39284c28-156f-445d-9951-9df6851c9487" containerName="extract-content" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188654 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8518501-8647-42d5-9e20-6d69f40f1e5d" containerName="registry-server" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188673 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fbea3c3-376b-48e1-a863-3eb8478a3974" containerName="registry-server" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188697 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="76949538-95e6-4f55-a9e3-3dcccd40580e" containerName="registry-server" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188711 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2b22bdb-961e-4920-841a-2d5f4e352c2a" containerName="pruner" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.188734 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="39284c28-156f-445d-9951-9df6851c9487" containerName="registry-server" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.189327 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.191057 4884 file.go:109] "Unable to process watch event" err="can't process config file \"/etc/kubernetes/manifests/kube-apiserver-pod.yaml\": /etc/kubernetes/manifests/kube-apiserver-pod.yaml: couldn't parse as pod(Object 'Kind' is missing in 'null'), please check config file" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.191296 4884 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.191975 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234" gracePeriod=15 Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.192006 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4" gracePeriod=15 Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.192045 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310" gracePeriod=15 Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.192089 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf" gracePeriod=15 Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.192091 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8" gracePeriod=15 Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.193482 4884 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.193718 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.193769 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.193791 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.193804 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.193822 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.193834 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.193854 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.193866 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.193884 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.193897 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.193918 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.193930 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.194133 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.194157 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.194172 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.194187 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.194199 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.194222 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 01:40:12 crc kubenswrapper[4884]: E1202 01:40:12.194372 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.194386 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.317364 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.317441 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.317475 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.317506 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.317547 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.317603 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.317633 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.317681 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418540 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418583 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418603 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418626 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418646 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418676 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418690 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418686 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418713 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418795 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418816 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418844 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418851 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418881 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418894 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.418916 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.486814 4884 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 02 01:40:12 crc kubenswrapper[4884]: I1202 01:40:12.486887 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 02 01:40:13 crc kubenswrapper[4884]: I1202 01:40:13.237214 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 02 01:40:13 crc kubenswrapper[4884]: I1202 01:40:13.239630 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 02 01:40:13 crc kubenswrapper[4884]: I1202 01:40:13.240842 4884 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4" exitCode=0 Dec 02 01:40:13 crc kubenswrapper[4884]: I1202 01:40:13.240884 4884 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf" exitCode=0 Dec 02 01:40:13 crc kubenswrapper[4884]: I1202 01:40:13.240902 4884 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310" exitCode=0 Dec 02 01:40:13 crc kubenswrapper[4884]: I1202 01:40:13.240916 4884 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8" exitCode=2 Dec 02 01:40:13 crc kubenswrapper[4884]: I1202 01:40:13.240950 4884 scope.go:117] "RemoveContainer" containerID="aabd1e74e47ae76062abacf60c3798a265511c04d0a59847061f23cdcd7e82db" Dec 02 01:40:13 crc kubenswrapper[4884]: I1202 01:40:13.245159 4884 generic.go:334] "Generic (PLEG): container finished" podID="ca937f60-dd07-4e98-a9aa-1ab893593165" containerID="05dc4b2aa5878107884bd06f3617013fc5917ed427e2ab6dec4aaecceb1a7dc9" exitCode=0 Dec 02 01:40:13 crc kubenswrapper[4884]: I1202 01:40:13.245219 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ca937f60-dd07-4e98-a9aa-1ab893593165","Type":"ContainerDied","Data":"05dc4b2aa5878107884bd06f3617013fc5917ed427e2ab6dec4aaecceb1a7dc9"} Dec 02 01:40:13 crc kubenswrapper[4884]: I1202 01:40:13.246970 4884 status_manager.go:851] "Failed to get status for pod" podUID="ca937f60-dd07-4e98-a9aa-1ab893593165" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:13 crc kubenswrapper[4884]: I1202 01:40:13.247622 4884 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:13 crc kubenswrapper[4884]: I1202 01:40:13.626886 4884 status_manager.go:851] "Failed to get status for pod" podUID="ca937f60-dd07-4e98-a9aa-1ab893593165" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:13 crc kubenswrapper[4884]: I1202 01:40:13.627458 4884 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.255348 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.594674 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.595725 4884 status_manager.go:851] "Failed to get status for pod" podUID="ca937f60-dd07-4e98-a9aa-1ab893593165" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.599131 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.600559 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.601031 4884 status_manager.go:851] "Failed to get status for pod" podUID="ca937f60-dd07-4e98-a9aa-1ab893593165" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.601321 4884 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.672634 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.672720 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ca937f60-dd07-4e98-a9aa-1ab893593165-kube-api-access\") pod \"ca937f60-dd07-4e98-a9aa-1ab893593165\" (UID: \"ca937f60-dd07-4e98-a9aa-1ab893593165\") " Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.672792 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.672820 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ca937f60-dd07-4e98-a9aa-1ab893593165-var-lock\") pod \"ca937f60-dd07-4e98-a9aa-1ab893593165\" (UID: \"ca937f60-dd07-4e98-a9aa-1ab893593165\") " Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.672860 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.672884 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ca937f60-dd07-4e98-a9aa-1ab893593165-kubelet-dir\") pod \"ca937f60-dd07-4e98-a9aa-1ab893593165\" (UID: \"ca937f60-dd07-4e98-a9aa-1ab893593165\") " Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.672923 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca937f60-dd07-4e98-a9aa-1ab893593165-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ca937f60-dd07-4e98-a9aa-1ab893593165" (UID: "ca937f60-dd07-4e98-a9aa-1ab893593165"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.672944 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.672997 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.673034 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca937f60-dd07-4e98-a9aa-1ab893593165-var-lock" (OuterVolumeSpecName: "var-lock") pod "ca937f60-dd07-4e98-a9aa-1ab893593165" (UID: "ca937f60-dd07-4e98-a9aa-1ab893593165"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.673094 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.673660 4884 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.673686 4884 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.673705 4884 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ca937f60-dd07-4e98-a9aa-1ab893593165-var-lock\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.673722 4884 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ca937f60-dd07-4e98-a9aa-1ab893593165-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.673739 4884 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.681906 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca937f60-dd07-4e98-a9aa-1ab893593165-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ca937f60-dd07-4e98-a9aa-1ab893593165" (UID: "ca937f60-dd07-4e98-a9aa-1ab893593165"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:40:14 crc kubenswrapper[4884]: I1202 01:40:14.775048 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ca937f60-dd07-4e98-a9aa-1ab893593165-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.273112 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.274649 4884 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234" exitCode=0 Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.274822 4884 scope.go:117] "RemoveContainer" containerID="b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.275026 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.279100 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"ca937f60-dd07-4e98-a9aa-1ab893593165","Type":"ContainerDied","Data":"914d31058d7b22435d96a6027e1e4dd84c2353d177e03a1e6ae9cb8eba2e995e"} Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.279165 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="914d31058d7b22435d96a6027e1e4dd84c2353d177e03a1e6ae9cb8eba2e995e" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.279476 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.308574 4884 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.309236 4884 status_manager.go:851] "Failed to get status for pod" podUID="ca937f60-dd07-4e98-a9aa-1ab893593165" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.309725 4884 status_manager.go:851] "Failed to get status for pod" podUID="ca937f60-dd07-4e98-a9aa-1ab893593165" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.310183 4884 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.313474 4884 scope.go:117] "RemoveContainer" containerID="5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.339336 4884 scope.go:117] "RemoveContainer" containerID="1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.366509 4884 scope.go:117] "RemoveContainer" containerID="8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.391193 4884 scope.go:117] "RemoveContainer" containerID="b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.428421 4884 scope.go:117] "RemoveContainer" containerID="d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.465968 4884 scope.go:117] "RemoveContainer" containerID="b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4" Dec 02 01:40:15 crc kubenswrapper[4884]: E1202 01:40:15.468915 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\": container with ID starting with b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4 not found: ID does not exist" containerID="b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.468990 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4"} err="failed to get container status \"b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\": rpc error: code = NotFound desc = could not find container \"b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4\": container with ID starting with b0006db3f33f3eec712dfdfcbaafff61ef4c170b38c33d1b96af7dca48f281a4 not found: ID does not exist" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.469431 4884 scope.go:117] "RemoveContainer" containerID="5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf" Dec 02 01:40:15 crc kubenswrapper[4884]: E1202 01:40:15.471385 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\": container with ID starting with 5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf not found: ID does not exist" containerID="5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.471505 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf"} err="failed to get container status \"5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\": rpc error: code = NotFound desc = could not find container \"5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf\": container with ID starting with 5223cfc16d49237242be3f667e8feafecfa2a6fe5ffec47abc8af7ebc62fb3cf not found: ID does not exist" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.471619 4884 scope.go:117] "RemoveContainer" containerID="1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310" Dec 02 01:40:15 crc kubenswrapper[4884]: E1202 01:40:15.473105 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\": container with ID starting with 1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310 not found: ID does not exist" containerID="1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.473199 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310"} err="failed to get container status \"1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\": rpc error: code = NotFound desc = could not find container \"1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310\": container with ID starting with 1b2a288f354a2d065737d95ae35bebd4ce2eabcaff4503a1fb150246006a6310 not found: ID does not exist" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.473279 4884 scope.go:117] "RemoveContainer" containerID="8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8" Dec 02 01:40:15 crc kubenswrapper[4884]: E1202 01:40:15.473970 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\": container with ID starting with 8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8 not found: ID does not exist" containerID="8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.474171 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8"} err="failed to get container status \"8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\": rpc error: code = NotFound desc = could not find container \"8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8\": container with ID starting with 8eb53e8ac97ec7ab2924b6fda731ed8779b22e1e4d77253573cedeb5af43e9f8 not found: ID does not exist" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.474256 4884 scope.go:117] "RemoveContainer" containerID="b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234" Dec 02 01:40:15 crc kubenswrapper[4884]: E1202 01:40:15.475036 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\": container with ID starting with b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234 not found: ID does not exist" containerID="b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.475089 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234"} err="failed to get container status \"b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\": rpc error: code = NotFound desc = could not find container \"b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234\": container with ID starting with b3a92573d4156a902b5dd0d8245c57316d096a21a7ec9d22c6d5c3aebb337234 not found: ID does not exist" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.475178 4884 scope.go:117] "RemoveContainer" containerID="d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7" Dec 02 01:40:15 crc kubenswrapper[4884]: E1202 01:40:15.475740 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\": container with ID starting with d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7 not found: ID does not exist" containerID="d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.475811 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7"} err="failed to get container status \"d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\": rpc error: code = NotFound desc = could not find container \"d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7\": container with ID starting with d9cd8db3d82811cd08d14dded8ef437d60252dbdcb3dff003d1bfd087efe59f7 not found: ID does not exist" Dec 02 01:40:15 crc kubenswrapper[4884]: I1202 01:40:15.622517 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 02 01:40:16 crc kubenswrapper[4884]: E1202 01:40:16.227554 4884 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:16 crc kubenswrapper[4884]: E1202 01:40:16.228588 4884 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:16 crc kubenswrapper[4884]: E1202 01:40:16.229316 4884 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:16 crc kubenswrapper[4884]: E1202 01:40:16.229848 4884 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:16 crc kubenswrapper[4884]: E1202 01:40:16.230433 4884 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:16 crc kubenswrapper[4884]: I1202 01:40:16.230490 4884 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 02 01:40:16 crc kubenswrapper[4884]: E1202 01:40:16.230947 4884 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="200ms" Dec 02 01:40:16 crc kubenswrapper[4884]: E1202 01:40:16.432027 4884 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="400ms" Dec 02 01:40:16 crc kubenswrapper[4884]: E1202 01:40:16.833717 4884 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="800ms" Dec 02 01:40:17 crc kubenswrapper[4884]: E1202 01:40:17.234556 4884 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.32:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:17 crc kubenswrapper[4884]: I1202 01:40:17.235311 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:17 crc kubenswrapper[4884]: E1202 01:40:17.283061 4884 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.32:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d4261e8b268a3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-02 01:40:17.282369699 +0000 UTC m=+233.958206633,LastTimestamp:2025-12-02 01:40:17.282369699 +0000 UTC m=+233.958206633,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 02 01:40:17 crc kubenswrapper[4884]: I1202 01:40:17.300816 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a4895c9a6cc856cc1d40e2c923ef6b39633032b22f2664d99baa3677057eab08"} Dec 02 01:40:17 crc kubenswrapper[4884]: E1202 01:40:17.635243 4884 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="1.6s" Dec 02 01:40:18 crc kubenswrapper[4884]: I1202 01:40:18.308478 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"ddc0fc83af2802b5f12c41da62e08d21218cc15970730b66d98dcef6566af326"} Dec 02 01:40:18 crc kubenswrapper[4884]: I1202 01:40:18.309390 4884 status_manager.go:851] "Failed to get status for pod" podUID="ca937f60-dd07-4e98-a9aa-1ab893593165" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:18 crc kubenswrapper[4884]: E1202 01:40:18.310115 4884 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.32:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:19 crc kubenswrapper[4884]: E1202 01:40:19.237060 4884 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="3.2s" Dec 02 01:40:19 crc kubenswrapper[4884]: E1202 01:40:19.318220 4884 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.32:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:40:22 crc kubenswrapper[4884]: E1202 01:40:22.438631 4884 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.32:6443: connect: connection refused" interval="6.4s" Dec 02 01:40:22 crc kubenswrapper[4884]: E1202 01:40:22.857344 4884 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.32:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187d4261e8b268a3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-02 01:40:17.282369699 +0000 UTC m=+233.958206633,LastTimestamp:2025-12-02 01:40:17.282369699 +0000 UTC m=+233.958206633,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 02 01:40:23 crc kubenswrapper[4884]: I1202 01:40:23.620942 4884 status_manager.go:851] "Failed to get status for pod" podUID="ca937f60-dd07-4e98-a9aa-1ab893593165" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:25 crc kubenswrapper[4884]: I1202 01:40:25.364301 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 02 01:40:25 crc kubenswrapper[4884]: I1202 01:40:25.364596 4884 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7" exitCode=1 Dec 02 01:40:25 crc kubenswrapper[4884]: I1202 01:40:25.364626 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7"} Dec 02 01:40:25 crc kubenswrapper[4884]: I1202 01:40:25.365120 4884 scope.go:117] "RemoveContainer" containerID="0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7" Dec 02 01:40:25 crc kubenswrapper[4884]: I1202 01:40:25.365519 4884 status_manager.go:851] "Failed to get status for pod" podUID="ca937f60-dd07-4e98-a9aa-1ab893593165" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:25 crc kubenswrapper[4884]: I1202 01:40:25.366143 4884 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:25 crc kubenswrapper[4884]: I1202 01:40:25.613296 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:25 crc kubenswrapper[4884]: I1202 01:40:25.614654 4884 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:25 crc kubenswrapper[4884]: I1202 01:40:25.615320 4884 status_manager.go:851] "Failed to get status for pod" podUID="ca937f60-dd07-4e98-a9aa-1ab893593165" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:25 crc kubenswrapper[4884]: I1202 01:40:25.650363 4884 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fdba8a22-47be-4221-b923-a41afee4989d" Dec 02 01:40:25 crc kubenswrapper[4884]: I1202 01:40:25.650415 4884 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fdba8a22-47be-4221-b923-a41afee4989d" Dec 02 01:40:25 crc kubenswrapper[4884]: E1202 01:40:25.651228 4884 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:25 crc kubenswrapper[4884]: I1202 01:40:25.652079 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:25 crc kubenswrapper[4884]: W1202 01:40:25.679543 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-cbfc2693bb544999b37c2ede4cdbef4b1b3f1bbc49427cfe50ff0c75a3b210b7 WatchSource:0}: Error finding container cbfc2693bb544999b37c2ede4cdbef4b1b3f1bbc49427cfe50ff0c75a3b210b7: Status 404 returned error can't find the container with id cbfc2693bb544999b37c2ede4cdbef4b1b3f1bbc49427cfe50ff0c75a3b210b7 Dec 02 01:40:26 crc kubenswrapper[4884]: I1202 01:40:26.373722 4884 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="8cc9c58a5a133159af273c969dc07f7ebabcd145461133e925f37e6a117abe78" exitCode=0 Dec 02 01:40:26 crc kubenswrapper[4884]: I1202 01:40:26.373858 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"8cc9c58a5a133159af273c969dc07f7ebabcd145461133e925f37e6a117abe78"} Dec 02 01:40:26 crc kubenswrapper[4884]: I1202 01:40:26.373926 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"cbfc2693bb544999b37c2ede4cdbef4b1b3f1bbc49427cfe50ff0c75a3b210b7"} Dec 02 01:40:26 crc kubenswrapper[4884]: I1202 01:40:26.374394 4884 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fdba8a22-47be-4221-b923-a41afee4989d" Dec 02 01:40:26 crc kubenswrapper[4884]: I1202 01:40:26.374427 4884 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fdba8a22-47be-4221-b923-a41afee4989d" Dec 02 01:40:26 crc kubenswrapper[4884]: I1202 01:40:26.375065 4884 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:26 crc kubenswrapper[4884]: E1202 01:40:26.375107 4884 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:26 crc kubenswrapper[4884]: I1202 01:40:26.375510 4884 status_manager.go:851] "Failed to get status for pod" podUID="ca937f60-dd07-4e98-a9aa-1ab893593165" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:26 crc kubenswrapper[4884]: I1202 01:40:26.380237 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 02 01:40:26 crc kubenswrapper[4884]: I1202 01:40:26.380316 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"88db83992afff1905ad9b38c18c32d592a05a71a3b8546c1b9070412d17f863b"} Dec 02 01:40:26 crc kubenswrapper[4884]: I1202 01:40:26.381470 4884 status_manager.go:851] "Failed to get status for pod" podUID="ca937f60-dd07-4e98-a9aa-1ab893593165" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:26 crc kubenswrapper[4884]: I1202 01:40:26.382022 4884 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.32:6443: connect: connection refused" Dec 02 01:40:27 crc kubenswrapper[4884]: I1202 01:40:27.389580 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5d3a9a5b6609c55c2fffa427a6d9f1920ed70217d534601ba6294d1b31c37ce3"} Dec 02 01:40:27 crc kubenswrapper[4884]: I1202 01:40:27.389982 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b34fbda91aa7c7a9ceb2315c0064b4bf45fa2a7112d32478f1126956249c49a2"} Dec 02 01:40:28 crc kubenswrapper[4884]: I1202 01:40:28.397987 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8c8bc15afb33c428457c080e2dd5d8a50ce78dc87ccacf75b28f3fc7aec55b42"} Dec 02 01:40:28 crc kubenswrapper[4884]: I1202 01:40:28.398268 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e63a29b30c86c775b9fc09a5ee1c1f5ac98a1c60d523434dbcb6fbb68fbc086d"} Dec 02 01:40:28 crc kubenswrapper[4884]: I1202 01:40:28.398282 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0638aa8e1b92a0756b71ff78649a0fe0c2913f7aedf3ebb21c600b96176c12b7"} Dec 02 01:40:28 crc kubenswrapper[4884]: I1202 01:40:28.398303 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:28 crc kubenswrapper[4884]: I1202 01:40:28.398342 4884 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fdba8a22-47be-4221-b923-a41afee4989d" Dec 02 01:40:28 crc kubenswrapper[4884]: I1202 01:40:28.398373 4884 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fdba8a22-47be-4221-b923-a41afee4989d" Dec 02 01:40:29 crc kubenswrapper[4884]: I1202 01:40:29.071637 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" podUID="0b33acfc-c22d-4d1f-a70f-410ed90474bb" containerName="oauth-openshift" containerID="cri-o://c5b646528bdf8be50f4595e62beed764eaeadefcb1cf90076d48db880c1685b9" gracePeriod=15 Dec 02 01:40:29 crc kubenswrapper[4884]: I1202 01:40:29.407088 4884 generic.go:334] "Generic (PLEG): container finished" podID="0b33acfc-c22d-4d1f-a70f-410ed90474bb" containerID="c5b646528bdf8be50f4595e62beed764eaeadefcb1cf90076d48db880c1685b9" exitCode=0 Dec 02 01:40:29 crc kubenswrapper[4884]: I1202 01:40:29.407162 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" event={"ID":"0b33acfc-c22d-4d1f-a70f-410ed90474bb","Type":"ContainerDied","Data":"c5b646528bdf8be50f4595e62beed764eaeadefcb1cf90076d48db880c1685b9"} Dec 02 01:40:29 crc kubenswrapper[4884]: I1202 01:40:29.542863 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:40:30 crc kubenswrapper[4884]: I1202 01:40:30.148659 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:40:30 crc kubenswrapper[4884]: I1202 01:40:30.415581 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" event={"ID":"0b33acfc-c22d-4d1f-a70f-410ed90474bb","Type":"ContainerDied","Data":"f23b4b28662d8772cf9741077b471b5d983027cbd37f0e72e22fe60e1be95ba2"} Dec 02 01:40:30 crc kubenswrapper[4884]: I1202 01:40:30.415677 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-wpxzm" Dec 02 01:40:30 crc kubenswrapper[4884]: I1202 01:40:30.415949 4884 scope.go:117] "RemoveContainer" containerID="c5b646528bdf8be50f4595e62beed764eaeadefcb1cf90076d48db880c1685b9" Dec 02 01:40:30 crc kubenswrapper[4884]: I1202 01:40:30.652583 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:30 crc kubenswrapper[4884]: I1202 01:40:30.652663 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:30 crc kubenswrapper[4884]: I1202 01:40:30.661186 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.413214 4884 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.439494 4884 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fdba8a22-47be-4221-b923-a41afee4989d" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.439535 4884 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fdba8a22-47be-4221-b923-a41afee4989d" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.446792 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.640066 4884 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="63bb2c71-e40a-4181-a852-cc6a42a25d7a" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.710289 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-provider-selection\") pod \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.710400 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0b33acfc-c22d-4d1f-a70f-410ed90474bb-audit-dir\") pod \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.710446 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-idp-0-file-data\") pod \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.710517 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-router-certs\") pod \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.710552 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0b33acfc-c22d-4d1f-a70f-410ed90474bb-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "0b33acfc-c22d-4d1f-a70f-410ed90474bb" (UID: "0b33acfc-c22d-4d1f-a70f-410ed90474bb"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.710575 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-serving-cert\") pod \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.710694 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-login\") pod \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.710765 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-audit-policies\") pod \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.710839 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-ocp-branding-template\") pod \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.710935 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-error\") pod \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.710996 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z495x\" (UniqueName: \"kubernetes.io/projected/0b33acfc-c22d-4d1f-a70f-410ed90474bb-kube-api-access-z495x\") pod \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.711040 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-service-ca\") pod \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.711091 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-session\") pod \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.711158 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-cliconfig\") pod \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.711205 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-trusted-ca-bundle\") pod \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\" (UID: \"0b33acfc-c22d-4d1f-a70f-410ed90474bb\") " Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.712151 4884 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0b33acfc-c22d-4d1f-a70f-410ed90474bb-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.712467 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "0b33acfc-c22d-4d1f-a70f-410ed90474bb" (UID: "0b33acfc-c22d-4d1f-a70f-410ed90474bb"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.712800 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "0b33acfc-c22d-4d1f-a70f-410ed90474bb" (UID: "0b33acfc-c22d-4d1f-a70f-410ed90474bb"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.713016 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "0b33acfc-c22d-4d1f-a70f-410ed90474bb" (UID: "0b33acfc-c22d-4d1f-a70f-410ed90474bb"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.716025 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "0b33acfc-c22d-4d1f-a70f-410ed90474bb" (UID: "0b33acfc-c22d-4d1f-a70f-410ed90474bb"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.720328 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "0b33acfc-c22d-4d1f-a70f-410ed90474bb" (UID: "0b33acfc-c22d-4d1f-a70f-410ed90474bb"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.722036 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "0b33acfc-c22d-4d1f-a70f-410ed90474bb" (UID: "0b33acfc-c22d-4d1f-a70f-410ed90474bb"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.722681 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "0b33acfc-c22d-4d1f-a70f-410ed90474bb" (UID: "0b33acfc-c22d-4d1f-a70f-410ed90474bb"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.723174 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b33acfc-c22d-4d1f-a70f-410ed90474bb-kube-api-access-z495x" (OuterVolumeSpecName: "kube-api-access-z495x") pod "0b33acfc-c22d-4d1f-a70f-410ed90474bb" (UID: "0b33acfc-c22d-4d1f-a70f-410ed90474bb"). InnerVolumeSpecName "kube-api-access-z495x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.723798 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "0b33acfc-c22d-4d1f-a70f-410ed90474bb" (UID: "0b33acfc-c22d-4d1f-a70f-410ed90474bb"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.723807 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "0b33acfc-c22d-4d1f-a70f-410ed90474bb" (UID: "0b33acfc-c22d-4d1f-a70f-410ed90474bb"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.723934 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "0b33acfc-c22d-4d1f-a70f-410ed90474bb" (UID: "0b33acfc-c22d-4d1f-a70f-410ed90474bb"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.724158 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "0b33acfc-c22d-4d1f-a70f-410ed90474bb" (UID: "0b33acfc-c22d-4d1f-a70f-410ed90474bb"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.729181 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "0b33acfc-c22d-4d1f-a70f-410ed90474bb" (UID: "0b33acfc-c22d-4d1f-a70f-410ed90474bb"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.813673 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.813787 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.813822 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.813852 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.813882 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.813911 4884 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.813936 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.813959 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.813980 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z495x\" (UniqueName: \"kubernetes.io/projected/0b33acfc-c22d-4d1f-a70f-410ed90474bb-kube-api-access-z495x\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.813999 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.814019 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.814038 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:33 crc kubenswrapper[4884]: I1202 01:40:33.814061 4884 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0b33acfc-c22d-4d1f-a70f-410ed90474bb-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:40:34 crc kubenswrapper[4884]: I1202 01:40:34.448108 4884 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fdba8a22-47be-4221-b923-a41afee4989d" Dec 02 01:40:34 crc kubenswrapper[4884]: I1202 01:40:34.448157 4884 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="fdba8a22-47be-4221-b923-a41afee4989d" Dec 02 01:40:34 crc kubenswrapper[4884]: I1202 01:40:34.450537 4884 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="63bb2c71-e40a-4181-a852-cc6a42a25d7a" Dec 02 01:40:34 crc kubenswrapper[4884]: E1202 01:40:34.456688 4884 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-idp-0-file-data\": Failed to watch *v1.Secret: unknown (get secrets)" logger="UnhandledError" Dec 02 01:40:34 crc kubenswrapper[4884]: I1202 01:40:34.994610 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:40:34 crc kubenswrapper[4884]: I1202 01:40:34.995592 4884 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 02 01:40:34 crc kubenswrapper[4884]: I1202 01:40:34.996284 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 02 01:40:42 crc kubenswrapper[4884]: I1202 01:40:42.916674 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 02 01:40:43 crc kubenswrapper[4884]: I1202 01:40:43.460772 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 02 01:40:43 crc kubenswrapper[4884]: I1202 01:40:43.498794 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 02 01:40:44 crc kubenswrapper[4884]: I1202 01:40:44.069181 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 02 01:40:44 crc kubenswrapper[4884]: I1202 01:40:44.272891 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 02 01:40:44 crc kubenswrapper[4884]: I1202 01:40:44.444177 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 02 01:40:44 crc kubenswrapper[4884]: I1202 01:40:44.702445 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 02 01:40:44 crc kubenswrapper[4884]: I1202 01:40:44.714115 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 02 01:40:44 crc kubenswrapper[4884]: I1202 01:40:44.806016 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 02 01:40:44 crc kubenswrapper[4884]: I1202 01:40:44.891190 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 02 01:40:44 crc kubenswrapper[4884]: I1202 01:40:44.926488 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 02 01:40:44 crc kubenswrapper[4884]: I1202 01:40:44.997081 4884 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 02 01:40:44 crc kubenswrapper[4884]: I1202 01:40:44.997249 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 02 01:40:45 crc kubenswrapper[4884]: I1202 01:40:45.049148 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 02 01:40:45 crc kubenswrapper[4884]: I1202 01:40:45.075885 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 02 01:40:45 crc kubenswrapper[4884]: I1202 01:40:45.132072 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 02 01:40:45 crc kubenswrapper[4884]: I1202 01:40:45.176923 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 02 01:40:45 crc kubenswrapper[4884]: I1202 01:40:45.329185 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 02 01:40:45 crc kubenswrapper[4884]: I1202 01:40:45.623466 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 02 01:40:45 crc kubenswrapper[4884]: I1202 01:40:45.889899 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 02 01:40:45 crc kubenswrapper[4884]: I1202 01:40:45.908298 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 02 01:40:45 crc kubenswrapper[4884]: I1202 01:40:45.941488 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 02 01:40:46 crc kubenswrapper[4884]: I1202 01:40:46.124124 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 02 01:40:46 crc kubenswrapper[4884]: I1202 01:40:46.191579 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 02 01:40:46 crc kubenswrapper[4884]: I1202 01:40:46.273508 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 02 01:40:46 crc kubenswrapper[4884]: I1202 01:40:46.294926 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 02 01:40:46 crc kubenswrapper[4884]: I1202 01:40:46.370053 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 02 01:40:46 crc kubenswrapper[4884]: I1202 01:40:46.394366 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 02 01:40:46 crc kubenswrapper[4884]: I1202 01:40:46.468368 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 02 01:40:46 crc kubenswrapper[4884]: I1202 01:40:46.512503 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 02 01:40:46 crc kubenswrapper[4884]: I1202 01:40:46.516865 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 02 01:40:46 crc kubenswrapper[4884]: I1202 01:40:46.709893 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 02 01:40:46 crc kubenswrapper[4884]: I1202 01:40:46.738660 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 02 01:40:46 crc kubenswrapper[4884]: I1202 01:40:46.821845 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 02 01:40:46 crc kubenswrapper[4884]: I1202 01:40:46.916137 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.012378 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.031973 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.035388 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.044024 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.085876 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.119493 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.160411 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.168538 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.327666 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.441809 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.473169 4884 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.570089 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.608033 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.675650 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.726578 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.873878 4884 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.880497 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-wpxzm","openshift-kube-apiserver/kube-apiserver-crc"] Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.880612 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.887717 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.901690 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.910386 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=14.910357347 podStartE2EDuration="14.910357347s" podCreationTimestamp="2025-12-02 01:40:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:40:47.905072438 +0000 UTC m=+264.580909322" watchObservedRunningTime="2025-12-02 01:40:47.910357347 +0000 UTC m=+264.586194271" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.935060 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.948309 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.981782 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 02 01:40:47 crc kubenswrapper[4884]: I1202 01:40:47.993550 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.056345 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.060801 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.213736 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.215670 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.220952 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.270423 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.503028 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.507423 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.542766 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.555493 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.577094 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.678602 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.719480 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.730917 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.747391 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.769647 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.854700 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.923735 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.929024 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.954320 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 02 01:40:48 crc kubenswrapper[4884]: I1202 01:40:48.983039 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 02 01:40:49 crc kubenswrapper[4884]: I1202 01:40:49.014801 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 02 01:40:49 crc kubenswrapper[4884]: I1202 01:40:49.044329 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 02 01:40:49 crc kubenswrapper[4884]: I1202 01:40:49.137356 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 02 01:40:49 crc kubenswrapper[4884]: I1202 01:40:49.314672 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 02 01:40:49 crc kubenswrapper[4884]: I1202 01:40:49.408573 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 02 01:40:49 crc kubenswrapper[4884]: I1202 01:40:49.493289 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 02 01:40:49 crc kubenswrapper[4884]: I1202 01:40:49.496348 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 02 01:40:49 crc kubenswrapper[4884]: I1202 01:40:49.552211 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 02 01:40:49 crc kubenswrapper[4884]: I1202 01:40:49.626354 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b33acfc-c22d-4d1f-a70f-410ed90474bb" path="/var/lib/kubelet/pods/0b33acfc-c22d-4d1f-a70f-410ed90474bb/volumes" Dec 02 01:40:49 crc kubenswrapper[4884]: I1202 01:40:49.747330 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 02 01:40:49 crc kubenswrapper[4884]: I1202 01:40:49.774720 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 02 01:40:49 crc kubenswrapper[4884]: I1202 01:40:49.787562 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 02 01:40:49 crc kubenswrapper[4884]: I1202 01:40:49.868723 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 02 01:40:49 crc kubenswrapper[4884]: I1202 01:40:49.896191 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 02 01:40:49 crc kubenswrapper[4884]: I1202 01:40:49.935738 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.045832 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.249599 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.282697 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.315823 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.458585 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-679fb67f4b-cvndp"] Dec 02 01:40:50 crc kubenswrapper[4884]: E1202 01:40:50.459436 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca937f60-dd07-4e98-a9aa-1ab893593165" containerName="installer" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.459466 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca937f60-dd07-4e98-a9aa-1ab893593165" containerName="installer" Dec 02 01:40:50 crc kubenswrapper[4884]: E1202 01:40:50.459503 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b33acfc-c22d-4d1f-a70f-410ed90474bb" containerName="oauth-openshift" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.459517 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b33acfc-c22d-4d1f-a70f-410ed90474bb" containerName="oauth-openshift" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.459695 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca937f60-dd07-4e98-a9aa-1ab893593165" containerName="installer" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.459722 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b33acfc-c22d-4d1f-a70f-410ed90474bb" containerName="oauth-openshift" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.460328 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.465461 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.466115 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.466669 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.467340 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.468720 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.469061 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.471151 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.471347 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.471526 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.471711 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.473038 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.476441 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.476790 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.483167 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.491495 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.497971 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.530934 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.545423 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4vh9\" (UniqueName: \"kubernetes.io/projected/3c67d09b-c7a4-421c-9959-1f95c911f2bd-kube-api-access-h4vh9\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.545503 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-service-ca\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.545535 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.545640 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-router-certs\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.545723 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.545820 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3c67d09b-c7a4-421c-9959-1f95c911f2bd-audit-dir\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.545894 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.545946 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-user-template-error\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.546000 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-user-template-login\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.546082 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.546134 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-session\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.546238 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.546317 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.546367 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3c67d09b-c7a4-421c-9959-1f95c911f2bd-audit-policies\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.647715 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.648198 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-session\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.648289 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.648346 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.648390 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3c67d09b-c7a4-421c-9959-1f95c911f2bd-audit-policies\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.648452 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4vh9\" (UniqueName: \"kubernetes.io/projected/3c67d09b-c7a4-421c-9959-1f95c911f2bd-kube-api-access-h4vh9\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.648508 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-service-ca\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.648543 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.648578 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-router-certs\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.648621 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.648696 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3c67d09b-c7a4-421c-9959-1f95c911f2bd-audit-dir\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.648737 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.648808 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-user-template-error\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.648849 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-user-template-login\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.649086 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.650340 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3c67d09b-c7a4-421c-9959-1f95c911f2bd-audit-dir\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.652253 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.652596 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3c67d09b-c7a4-421c-9959-1f95c911f2bd-audit-policies\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.656493 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-service-ca\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.657576 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.658860 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.660273 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-user-template-error\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.660795 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.661629 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.662479 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-user-template-login\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.667365 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-router-certs\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.669164 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3c67d09b-c7a4-421c-9959-1f95c911f2bd-v4-0-config-system-session\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.685253 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4vh9\" (UniqueName: \"kubernetes.io/projected/3c67d09b-c7a4-421c-9959-1f95c911f2bd-kube-api-access-h4vh9\") pod \"oauth-openshift-679fb67f4b-cvndp\" (UID: \"3c67d09b-c7a4-421c-9959-1f95c911f2bd\") " pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.725318 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.726473 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.728119 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.780049 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.791500 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.893647 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 02 01:40:50 crc kubenswrapper[4884]: I1202 01:40:50.936316 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.021688 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.068794 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.084937 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.150040 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.236037 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.243791 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.329518 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.343666 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.350235 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.414444 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.494800 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.652637 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.835509 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.877267 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.906046 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.922720 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 02 01:40:51 crc kubenswrapper[4884]: I1202 01:40:51.949587 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.025606 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.120562 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.135860 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.155461 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.348458 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.359653 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.377695 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.407324 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.504990 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.537687 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.541800 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.624233 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.708175 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.717016 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.741332 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.780247 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.848500 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.960849 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 02 01:40:52 crc kubenswrapper[4884]: I1202 01:40:52.991917 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.000786 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.040110 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.118286 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.173806 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.180190 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.181144 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.182915 4884 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.185634 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.204166 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.259991 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.310459 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.406865 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.489942 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.512172 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.514086 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.575885 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.669896 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.679696 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.690453 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.748957 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.850837 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.852856 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.868102 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 02 01:40:53 crc kubenswrapper[4884]: I1202 01:40:53.899007 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.043971 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.060420 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.106147 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.111849 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.176468 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.297317 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.306861 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.313915 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.326141 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.338520 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.464910 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.491718 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.520526 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.540681 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.579204 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.579206 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.624205 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.658033 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.821023 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.899065 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.976858 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.995248 4884 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.995311 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.995378 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.996249 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"88db83992afff1905ad9b38c18c32d592a05a71a3b8546c1b9070412d17f863b"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 02 01:40:54 crc kubenswrapper[4884]: I1202 01:40:54.996444 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://88db83992afff1905ad9b38c18c32d592a05a71a3b8546c1b9070412d17f863b" gracePeriod=30 Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.271537 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.303143 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.383573 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.455575 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.471008 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.478238 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.517352 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.540298 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.561074 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.578420 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.747540 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.760593 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.784825 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.856464 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.897934 4884 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.972471 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 02 01:40:55 crc kubenswrapper[4884]: I1202 01:40:55.976735 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.002571 4884 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.002884 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://ddc0fc83af2802b5f12c41da62e08d21218cc15970730b66d98dcef6566af326" gracePeriod=5 Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.070344 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.100928 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.136131 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.167823 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.217726 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.248312 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.260651 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.344994 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.364213 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.400401 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.491187 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.553569 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.721027 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 02 01:40:56 crc kubenswrapper[4884]: I1202 01:40:56.853521 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 02 01:40:57 crc kubenswrapper[4884]: I1202 01:40:57.074476 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 02 01:40:57 crc kubenswrapper[4884]: I1202 01:40:57.103236 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 02 01:40:57 crc kubenswrapper[4884]: I1202 01:40:57.231332 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 02 01:40:57 crc kubenswrapper[4884]: I1202 01:40:57.446438 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 02 01:40:57 crc kubenswrapper[4884]: I1202 01:40:57.458443 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 02 01:40:57 crc kubenswrapper[4884]: I1202 01:40:57.601073 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 02 01:40:57 crc kubenswrapper[4884]: I1202 01:40:57.780760 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 02 01:40:57 crc kubenswrapper[4884]: I1202 01:40:57.848775 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 02 01:40:57 crc kubenswrapper[4884]: I1202 01:40:57.863578 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 02 01:40:57 crc kubenswrapper[4884]: I1202 01:40:57.911781 4884 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 02 01:40:58 crc kubenswrapper[4884]: I1202 01:40:58.124791 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 02 01:40:58 crc kubenswrapper[4884]: I1202 01:40:58.273778 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 02 01:40:58 crc kubenswrapper[4884]: I1202 01:40:58.507482 4884 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 02 01:40:58 crc kubenswrapper[4884]: I1202 01:40:58.650727 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 02 01:40:58 crc kubenswrapper[4884]: I1202 01:40:58.653144 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 02 01:40:58 crc kubenswrapper[4884]: I1202 01:40:58.875371 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 02 01:40:58 crc kubenswrapper[4884]: I1202 01:40:58.891996 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 02 01:40:58 crc kubenswrapper[4884]: I1202 01:40:58.956653 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 02 01:40:58 crc kubenswrapper[4884]: I1202 01:40:58.975097 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 02 01:40:59 crc kubenswrapper[4884]: I1202 01:40:59.097982 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 02 01:40:59 crc kubenswrapper[4884]: I1202 01:40:59.159490 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-679fb67f4b-cvndp"] Dec 02 01:40:59 crc kubenswrapper[4884]: I1202 01:40:59.390967 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-679fb67f4b-cvndp"] Dec 02 01:40:59 crc kubenswrapper[4884]: I1202 01:40:59.621911 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" event={"ID":"3c67d09b-c7a4-421c-9959-1f95c911f2bd","Type":"ContainerStarted","Data":"a4c6e87f7e0cb670f45ef6b29c9506725fbd59d3259f6bb225e62524781bb4f1"} Dec 02 01:40:59 crc kubenswrapper[4884]: I1202 01:40:59.878145 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 02 01:41:00 crc kubenswrapper[4884]: I1202 01:41:00.428386 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 02 01:41:00 crc kubenswrapper[4884]: I1202 01:41:00.632868 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" event={"ID":"3c67d09b-c7a4-421c-9959-1f95c911f2bd","Type":"ContainerStarted","Data":"9412038fed2809595840a5291baec9396590f07f9c03c354f87da5c0db155789"} Dec 02 01:41:00 crc kubenswrapper[4884]: I1202 01:41:00.634253 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:41:00 crc kubenswrapper[4884]: I1202 01:41:00.645114 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" Dec 02 01:41:00 crc kubenswrapper[4884]: I1202 01:41:00.665008 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-679fb67f4b-cvndp" podStartSLOduration=56.66498534 podStartE2EDuration="56.66498534s" podCreationTimestamp="2025-12-02 01:40:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:41:00.664319494 +0000 UTC m=+277.340156428" watchObservedRunningTime="2025-12-02 01:41:00.66498534 +0000 UTC m=+277.340822264" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.066343 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.066523 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.613697 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.614038 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.641832 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.641911 4884 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="ddc0fc83af2802b5f12c41da62e08d21218cc15970730b66d98dcef6566af326" exitCode=137 Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.642000 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.642037 4884 scope.go:117] "RemoveContainer" containerID="ddc0fc83af2802b5f12c41da62e08d21218cc15970730b66d98dcef6566af326" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.663926 4884 scope.go:117] "RemoveContainer" containerID="ddc0fc83af2802b5f12c41da62e08d21218cc15970730b66d98dcef6566af326" Dec 02 01:41:01 crc kubenswrapper[4884]: E1202 01:41:01.664405 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddc0fc83af2802b5f12c41da62e08d21218cc15970730b66d98dcef6566af326\": container with ID starting with ddc0fc83af2802b5f12c41da62e08d21218cc15970730b66d98dcef6566af326 not found: ID does not exist" containerID="ddc0fc83af2802b5f12c41da62e08d21218cc15970730b66d98dcef6566af326" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.664468 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddc0fc83af2802b5f12c41da62e08d21218cc15970730b66d98dcef6566af326"} err="failed to get container status \"ddc0fc83af2802b5f12c41da62e08d21218cc15970730b66d98dcef6566af326\": rpc error: code = NotFound desc = could not find container \"ddc0fc83af2802b5f12c41da62e08d21218cc15970730b66d98dcef6566af326\": container with ID starting with ddc0fc83af2802b5f12c41da62e08d21218cc15970730b66d98dcef6566af326 not found: ID does not exist" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.713502 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.713573 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.713618 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.713610 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.713780 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.713979 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.714006 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.714033 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.714103 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.718240 4884 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.718289 4884 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.718310 4884 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.718330 4884 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.726425 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:41:01 crc kubenswrapper[4884]: I1202 01:41:01.820159 4884 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:03 crc kubenswrapper[4884]: I1202 01:41:03.626770 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 02 01:41:19 crc kubenswrapper[4884]: I1202 01:41:19.771188 4884 generic.go:334] "Generic (PLEG): container finished" podID="db61ff51-5c14-43bd-9042-cf7447ab7a3d" containerID="eb95cf030ae39b8d31a3de3fe32701ef5083b725345bb12e0d7bab83ddb770e1" exitCode=0 Dec 02 01:41:19 crc kubenswrapper[4884]: I1202 01:41:19.771246 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" event={"ID":"db61ff51-5c14-43bd-9042-cf7447ab7a3d","Type":"ContainerDied","Data":"eb95cf030ae39b8d31a3de3fe32701ef5083b725345bb12e0d7bab83ddb770e1"} Dec 02 01:41:19 crc kubenswrapper[4884]: I1202 01:41:19.772705 4884 scope.go:117] "RemoveContainer" containerID="eb95cf030ae39b8d31a3de3fe32701ef5083b725345bb12e0d7bab83ddb770e1" Dec 02 01:41:20 crc kubenswrapper[4884]: I1202 01:41:20.780523 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" event={"ID":"db61ff51-5c14-43bd-9042-cf7447ab7a3d","Type":"ContainerStarted","Data":"e9b569ae157f96a02bdcfa134878cb182f1cbc0c5e7e423023b35ceb151dfa09"} Dec 02 01:41:20 crc kubenswrapper[4884]: I1202 01:41:20.781351 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:41:20 crc kubenswrapper[4884]: I1202 01:41:20.783543 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:41:25 crc kubenswrapper[4884]: I1202 01:41:25.815351 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 02 01:41:25 crc kubenswrapper[4884]: I1202 01:41:25.817692 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 02 01:41:25 crc kubenswrapper[4884]: I1202 01:41:25.817771 4884 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="88db83992afff1905ad9b38c18c32d592a05a71a3b8546c1b9070412d17f863b" exitCode=137 Dec 02 01:41:25 crc kubenswrapper[4884]: I1202 01:41:25.817805 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"88db83992afff1905ad9b38c18c32d592a05a71a3b8546c1b9070412d17f863b"} Dec 02 01:41:25 crc kubenswrapper[4884]: I1202 01:41:25.817842 4884 scope.go:117] "RemoveContainer" containerID="0b9d1689843d94f2c22a1e843b5eea0546ae7711161020fc210a7339c28c8fa7" Dec 02 01:41:26 crc kubenswrapper[4884]: I1202 01:41:26.828148 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 02 01:41:26 crc kubenswrapper[4884]: I1202 01:41:26.829932 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fb12a356ced87de357cf81d79e7d07850aaf0e751f89f85b521aed221f0f0a83"} Dec 02 01:41:30 crc kubenswrapper[4884]: I1202 01:41:30.148241 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:41:34 crc kubenswrapper[4884]: I1202 01:41:34.995330 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:41:35 crc kubenswrapper[4884]: I1202 01:41:35.001322 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:41:35 crc kubenswrapper[4884]: I1202 01:41:35.900074 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 02 01:41:45 crc kubenswrapper[4884]: I1202 01:41:45.676944 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n757x"] Dec 02 01:41:45 crc kubenswrapper[4884]: I1202 01:41:45.677701 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" podUID="ccc071ec-9724-41f2-a654-625aae8bd840" containerName="controller-manager" containerID="cri-o://a62c28c965d6f35f3bdb46093015ed92a0f4a6b9cbfa286adb5dd37ace0a5712" gracePeriod=30 Dec 02 01:41:45 crc kubenswrapper[4884]: I1202 01:41:45.683733 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg"] Dec 02 01:41:45 crc kubenswrapper[4884]: I1202 01:41:45.683976 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" podUID="041d3d0d-70e0-4249-a7e2-63958513ba13" containerName="route-controller-manager" containerID="cri-o://a05f48787a2ffdedc90317f50de38683b6227dbe3bd576473d976566950174a0" gracePeriod=30 Dec 02 01:41:45 crc kubenswrapper[4884]: I1202 01:41:45.966949 4884 generic.go:334] "Generic (PLEG): container finished" podID="ccc071ec-9724-41f2-a654-625aae8bd840" containerID="a62c28c965d6f35f3bdb46093015ed92a0f4a6b9cbfa286adb5dd37ace0a5712" exitCode=0 Dec 02 01:41:45 crc kubenswrapper[4884]: I1202 01:41:45.967005 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" event={"ID":"ccc071ec-9724-41f2-a654-625aae8bd840","Type":"ContainerDied","Data":"a62c28c965d6f35f3bdb46093015ed92a0f4a6b9cbfa286adb5dd37ace0a5712"} Dec 02 01:41:45 crc kubenswrapper[4884]: I1202 01:41:45.969359 4884 generic.go:334] "Generic (PLEG): container finished" podID="041d3d0d-70e0-4249-a7e2-63958513ba13" containerID="a05f48787a2ffdedc90317f50de38683b6227dbe3bd576473d976566950174a0" exitCode=0 Dec 02 01:41:45 crc kubenswrapper[4884]: I1202 01:41:45.969393 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" event={"ID":"041d3d0d-70e0-4249-a7e2-63958513ba13","Type":"ContainerDied","Data":"a05f48787a2ffdedc90317f50de38683b6227dbe3bd576473d976566950174a0"} Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.144211 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.194430 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.319254 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/041d3d0d-70e0-4249-a7e2-63958513ba13-config\") pod \"041d3d0d-70e0-4249-a7e2-63958513ba13\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.319345 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-config\") pod \"ccc071ec-9724-41f2-a654-625aae8bd840\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.319376 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7v5b\" (UniqueName: \"kubernetes.io/projected/041d3d0d-70e0-4249-a7e2-63958513ba13-kube-api-access-q7v5b\") pod \"041d3d0d-70e0-4249-a7e2-63958513ba13\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.319394 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/041d3d0d-70e0-4249-a7e2-63958513ba13-serving-cert\") pod \"041d3d0d-70e0-4249-a7e2-63958513ba13\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.319416 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46zml\" (UniqueName: \"kubernetes.io/projected/ccc071ec-9724-41f2-a654-625aae8bd840-kube-api-access-46zml\") pod \"ccc071ec-9724-41f2-a654-625aae8bd840\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.319448 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-proxy-ca-bundles\") pod \"ccc071ec-9724-41f2-a654-625aae8bd840\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.319513 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-client-ca\") pod \"ccc071ec-9724-41f2-a654-625aae8bd840\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.319541 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccc071ec-9724-41f2-a654-625aae8bd840-serving-cert\") pod \"ccc071ec-9724-41f2-a654-625aae8bd840\" (UID: \"ccc071ec-9724-41f2-a654-625aae8bd840\") " Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.319561 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/041d3d0d-70e0-4249-a7e2-63958513ba13-client-ca\") pod \"041d3d0d-70e0-4249-a7e2-63958513ba13\" (UID: \"041d3d0d-70e0-4249-a7e2-63958513ba13\") " Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.320298 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/041d3d0d-70e0-4249-a7e2-63958513ba13-client-ca" (OuterVolumeSpecName: "client-ca") pod "041d3d0d-70e0-4249-a7e2-63958513ba13" (UID: "041d3d0d-70e0-4249-a7e2-63958513ba13"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.320616 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/041d3d0d-70e0-4249-a7e2-63958513ba13-config" (OuterVolumeSpecName: "config") pod "041d3d0d-70e0-4249-a7e2-63958513ba13" (UID: "041d3d0d-70e0-4249-a7e2-63958513ba13"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.321376 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-config" (OuterVolumeSpecName: "config") pod "ccc071ec-9724-41f2-a654-625aae8bd840" (UID: "ccc071ec-9724-41f2-a654-625aae8bd840"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.321473 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ccc071ec-9724-41f2-a654-625aae8bd840" (UID: "ccc071ec-9724-41f2-a654-625aae8bd840"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.321633 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-client-ca" (OuterVolumeSpecName: "client-ca") pod "ccc071ec-9724-41f2-a654-625aae8bd840" (UID: "ccc071ec-9724-41f2-a654-625aae8bd840"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.327083 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccc071ec-9724-41f2-a654-625aae8bd840-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ccc071ec-9724-41f2-a654-625aae8bd840" (UID: "ccc071ec-9724-41f2-a654-625aae8bd840"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.327362 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/041d3d0d-70e0-4249-a7e2-63958513ba13-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "041d3d0d-70e0-4249-a7e2-63958513ba13" (UID: "041d3d0d-70e0-4249-a7e2-63958513ba13"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.327393 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccc071ec-9724-41f2-a654-625aae8bd840-kube-api-access-46zml" (OuterVolumeSpecName: "kube-api-access-46zml") pod "ccc071ec-9724-41f2-a654-625aae8bd840" (UID: "ccc071ec-9724-41f2-a654-625aae8bd840"). InnerVolumeSpecName "kube-api-access-46zml". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.327441 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/041d3d0d-70e0-4249-a7e2-63958513ba13-kube-api-access-q7v5b" (OuterVolumeSpecName: "kube-api-access-q7v5b") pod "041d3d0d-70e0-4249-a7e2-63958513ba13" (UID: "041d3d0d-70e0-4249-a7e2-63958513ba13"). InnerVolumeSpecName "kube-api-access-q7v5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.421053 4884 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.421111 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ccc071ec-9724-41f2-a654-625aae8bd840-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.421130 4884 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/041d3d0d-70e0-4249-a7e2-63958513ba13-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.421146 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/041d3d0d-70e0-4249-a7e2-63958513ba13-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.421164 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.421182 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7v5b\" (UniqueName: \"kubernetes.io/projected/041d3d0d-70e0-4249-a7e2-63958513ba13-kube-api-access-q7v5b\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.421200 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/041d3d0d-70e0-4249-a7e2-63958513ba13-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.421219 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46zml\" (UniqueName: \"kubernetes.io/projected/ccc071ec-9724-41f2-a654-625aae8bd840-kube-api-access-46zml\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.421237 4884 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ccc071ec-9724-41f2-a654-625aae8bd840-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.793142 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx"] Dec 02 01:41:46 crc kubenswrapper[4884]: E1202 01:41:46.793719 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc071ec-9724-41f2-a654-625aae8bd840" containerName="controller-manager" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.793768 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc071ec-9724-41f2-a654-625aae8bd840" containerName="controller-manager" Dec 02 01:41:46 crc kubenswrapper[4884]: E1202 01:41:46.793802 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="041d3d0d-70e0-4249-a7e2-63958513ba13" containerName="route-controller-manager" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.793814 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="041d3d0d-70e0-4249-a7e2-63958513ba13" containerName="route-controller-manager" Dec 02 01:41:46 crc kubenswrapper[4884]: E1202 01:41:46.793833 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.793845 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.794025 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccc071ec-9724-41f2-a654-625aae8bd840" containerName="controller-manager" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.794060 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.794077 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="041d3d0d-70e0-4249-a7e2-63958513ba13" containerName="route-controller-manager" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.794909 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.805243 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-66f89d6466-dr6vt"] Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.806405 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.810053 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx"] Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.841562 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66f89d6466-dr6vt"] Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.929931 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-client-ca\") pod \"controller-manager-66f89d6466-dr6vt\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.929979 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7663a79d-5498-4459-8838-61d64eeb3d6d-serving-cert\") pod \"route-controller-manager-577fbf7ff6-86qqx\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.930018 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-config\") pod \"controller-manager-66f89d6466-dr6vt\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.930042 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7663a79d-5498-4459-8838-61d64eeb3d6d-config\") pod \"route-controller-manager-577fbf7ff6-86qqx\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.930061 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hvdw\" (UniqueName: \"kubernetes.io/projected/7663a79d-5498-4459-8838-61d64eeb3d6d-kube-api-access-4hvdw\") pod \"route-controller-manager-577fbf7ff6-86qqx\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.930078 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw9z4\" (UniqueName: \"kubernetes.io/projected/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-kube-api-access-rw9z4\") pod \"controller-manager-66f89d6466-dr6vt\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.930099 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7663a79d-5498-4459-8838-61d64eeb3d6d-client-ca\") pod \"route-controller-manager-577fbf7ff6-86qqx\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.930125 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-serving-cert\") pod \"controller-manager-66f89d6466-dr6vt\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.930413 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-proxy-ca-bundles\") pod \"controller-manager-66f89d6466-dr6vt\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.979042 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" event={"ID":"ccc071ec-9724-41f2-a654-625aae8bd840","Type":"ContainerDied","Data":"1bb44e4774c1bd8bf515059f82c105dfd3ee5eb9d4df49ed51d44409fa2ae6a4"} Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.979094 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-n757x" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.979126 4884 scope.go:117] "RemoveContainer" containerID="a62c28c965d6f35f3bdb46093015ed92a0f4a6b9cbfa286adb5dd37ace0a5712" Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.983994 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" event={"ID":"041d3d0d-70e0-4249-a7e2-63958513ba13","Type":"ContainerDied","Data":"2f34b84d127592f58fe99f02094b29e5bcdf83bf6f1147b79a9dfba1b4b49c2d"} Dec 02 01:41:46 crc kubenswrapper[4884]: I1202 01:41:46.984063 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.005811 4884 scope.go:117] "RemoveContainer" containerID="a05f48787a2ffdedc90317f50de38683b6227dbe3bd576473d976566950174a0" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.021255 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg"] Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.025383 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lqhjg"] Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.036343 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-proxy-ca-bundles\") pod \"controller-manager-66f89d6466-dr6vt\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.037495 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-client-ca\") pod \"controller-manager-66f89d6466-dr6vt\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.038315 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-proxy-ca-bundles\") pod \"controller-manager-66f89d6466-dr6vt\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.038346 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-client-ca\") pod \"controller-manager-66f89d6466-dr6vt\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.038367 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7663a79d-5498-4459-8838-61d64eeb3d6d-serving-cert\") pod \"route-controller-manager-577fbf7ff6-86qqx\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.038396 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-config\") pod \"controller-manager-66f89d6466-dr6vt\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.038414 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7663a79d-5498-4459-8838-61d64eeb3d6d-config\") pod \"route-controller-manager-577fbf7ff6-86qqx\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.038432 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hvdw\" (UniqueName: \"kubernetes.io/projected/7663a79d-5498-4459-8838-61d64eeb3d6d-kube-api-access-4hvdw\") pod \"route-controller-manager-577fbf7ff6-86qqx\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.038449 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw9z4\" (UniqueName: \"kubernetes.io/projected/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-kube-api-access-rw9z4\") pod \"controller-manager-66f89d6466-dr6vt\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.038465 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7663a79d-5498-4459-8838-61d64eeb3d6d-client-ca\") pod \"route-controller-manager-577fbf7ff6-86qqx\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.038488 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-serving-cert\") pod \"controller-manager-66f89d6466-dr6vt\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.044519 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7663a79d-5498-4459-8838-61d64eeb3d6d-config\") pod \"route-controller-manager-577fbf7ff6-86qqx\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.045789 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7663a79d-5498-4459-8838-61d64eeb3d6d-client-ca\") pod \"route-controller-manager-577fbf7ff6-86qqx\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.046019 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-config\") pod \"controller-manager-66f89d6466-dr6vt\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.050408 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n757x"] Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.051182 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7663a79d-5498-4459-8838-61d64eeb3d6d-serving-cert\") pod \"route-controller-manager-577fbf7ff6-86qqx\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.051418 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-serving-cert\") pod \"controller-manager-66f89d6466-dr6vt\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.056965 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-n757x"] Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.067923 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw9z4\" (UniqueName: \"kubernetes.io/projected/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-kube-api-access-rw9z4\") pod \"controller-manager-66f89d6466-dr6vt\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.078926 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hvdw\" (UniqueName: \"kubernetes.io/projected/7663a79d-5498-4459-8838-61d64eeb3d6d-kube-api-access-4hvdw\") pod \"route-controller-manager-577fbf7ff6-86qqx\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.114980 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.170932 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.401536 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx"] Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.489855 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66f89d6466-dr6vt"] Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.629298 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="041d3d0d-70e0-4249-a7e2-63958513ba13" path="/var/lib/kubelet/pods/041d3d0d-70e0-4249-a7e2-63958513ba13/volumes" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.630377 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccc071ec-9724-41f2-a654-625aae8bd840" path="/var/lib/kubelet/pods/ccc071ec-9724-41f2-a654-625aae8bd840/volumes" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.990968 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" event={"ID":"7663a79d-5498-4459-8838-61d64eeb3d6d","Type":"ContainerStarted","Data":"f68d77e4017e15dd19945e59720089116d7f45319bb351917853dda739ae5728"} Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.991508 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" event={"ID":"7663a79d-5498-4459-8838-61d64eeb3d6d","Type":"ContainerStarted","Data":"b064775c7c85f414d29d5f00d514b719130f5688b7e6629fe826e42062836d82"} Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.991526 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.992486 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" event={"ID":"232b9fb4-1281-4f12-ab4e-8b00aa0d4002","Type":"ContainerStarted","Data":"b38d6faf09251e97809f01685c0c41d439d963eb6868838a8b0d8de4aa86193f"} Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.992510 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" event={"ID":"232b9fb4-1281-4f12-ab4e-8b00aa0d4002","Type":"ContainerStarted","Data":"7b7ea84d9c3fc0cc820bd1791d87e10511203e838e6cbaa7d549552313709261"} Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.992675 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.997031 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:41:47 crc kubenswrapper[4884]: I1202 01:41:47.997366 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:48 crc kubenswrapper[4884]: I1202 01:41:48.015119 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" podStartSLOduration=3.015090474 podStartE2EDuration="3.015090474s" podCreationTimestamp="2025-12-02 01:41:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:41:48.010016911 +0000 UTC m=+324.685853795" watchObservedRunningTime="2025-12-02 01:41:48.015090474 +0000 UTC m=+324.690927358" Dec 02 01:41:48 crc kubenswrapper[4884]: I1202 01:41:48.030603 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" podStartSLOduration=3.030577881 podStartE2EDuration="3.030577881s" podCreationTimestamp="2025-12-02 01:41:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:41:48.029889514 +0000 UTC m=+324.705726398" watchObservedRunningTime="2025-12-02 01:41:48.030577881 +0000 UTC m=+324.706414765" Dec 02 01:41:48 crc kubenswrapper[4884]: I1202 01:41:48.471560 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx"] Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.017038 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" podUID="7663a79d-5498-4459-8838-61d64eeb3d6d" containerName="route-controller-manager" containerID="cri-o://f68d77e4017e15dd19945e59720089116d7f45319bb351917853dda739ae5728" gracePeriod=30 Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.462380 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.486571 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt"] Dec 02 01:41:50 crc kubenswrapper[4884]: E1202 01:41:50.486880 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7663a79d-5498-4459-8838-61d64eeb3d6d" containerName="route-controller-manager" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.486897 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="7663a79d-5498-4459-8838-61d64eeb3d6d" containerName="route-controller-manager" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.487016 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="7663a79d-5498-4459-8838-61d64eeb3d6d" containerName="route-controller-manager" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.487408 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.513671 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt"] Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.609875 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7663a79d-5498-4459-8838-61d64eeb3d6d-serving-cert\") pod \"7663a79d-5498-4459-8838-61d64eeb3d6d\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.610023 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hvdw\" (UniqueName: \"kubernetes.io/projected/7663a79d-5498-4459-8838-61d64eeb3d6d-kube-api-access-4hvdw\") pod \"7663a79d-5498-4459-8838-61d64eeb3d6d\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.610072 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7663a79d-5498-4459-8838-61d64eeb3d6d-config\") pod \"7663a79d-5498-4459-8838-61d64eeb3d6d\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.610107 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7663a79d-5498-4459-8838-61d64eeb3d6d-client-ca\") pod \"7663a79d-5498-4459-8838-61d64eeb3d6d\" (UID: \"7663a79d-5498-4459-8838-61d64eeb3d6d\") " Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.610308 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf669\" (UniqueName: \"kubernetes.io/projected/8512e92b-437e-4d32-af6a-c3213841c5f1-kube-api-access-lf669\") pod \"route-controller-manager-6758cd79b9-jhxvt\" (UID: \"8512e92b-437e-4d32-af6a-c3213841c5f1\") " pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.610367 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8512e92b-437e-4d32-af6a-c3213841c5f1-serving-cert\") pod \"route-controller-manager-6758cd79b9-jhxvt\" (UID: \"8512e92b-437e-4d32-af6a-c3213841c5f1\") " pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.610416 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8512e92b-437e-4d32-af6a-c3213841c5f1-client-ca\") pod \"route-controller-manager-6758cd79b9-jhxvt\" (UID: \"8512e92b-437e-4d32-af6a-c3213841c5f1\") " pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.610487 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8512e92b-437e-4d32-af6a-c3213841c5f1-config\") pod \"route-controller-manager-6758cd79b9-jhxvt\" (UID: \"8512e92b-437e-4d32-af6a-c3213841c5f1\") " pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.610845 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7663a79d-5498-4459-8838-61d64eeb3d6d-client-ca" (OuterVolumeSpecName: "client-ca") pod "7663a79d-5498-4459-8838-61d64eeb3d6d" (UID: "7663a79d-5498-4459-8838-61d64eeb3d6d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.610923 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7663a79d-5498-4459-8838-61d64eeb3d6d-config" (OuterVolumeSpecName: "config") pod "7663a79d-5498-4459-8838-61d64eeb3d6d" (UID: "7663a79d-5498-4459-8838-61d64eeb3d6d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.633480 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7663a79d-5498-4459-8838-61d64eeb3d6d-kube-api-access-4hvdw" (OuterVolumeSpecName: "kube-api-access-4hvdw") pod "7663a79d-5498-4459-8838-61d64eeb3d6d" (UID: "7663a79d-5498-4459-8838-61d64eeb3d6d"). InnerVolumeSpecName "kube-api-access-4hvdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.633475 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7663a79d-5498-4459-8838-61d64eeb3d6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7663a79d-5498-4459-8838-61d64eeb3d6d" (UID: "7663a79d-5498-4459-8838-61d64eeb3d6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.711731 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8512e92b-437e-4d32-af6a-c3213841c5f1-serving-cert\") pod \"route-controller-manager-6758cd79b9-jhxvt\" (UID: \"8512e92b-437e-4d32-af6a-c3213841c5f1\") " pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.711913 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8512e92b-437e-4d32-af6a-c3213841c5f1-client-ca\") pod \"route-controller-manager-6758cd79b9-jhxvt\" (UID: \"8512e92b-437e-4d32-af6a-c3213841c5f1\") " pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.711944 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8512e92b-437e-4d32-af6a-c3213841c5f1-config\") pod \"route-controller-manager-6758cd79b9-jhxvt\" (UID: \"8512e92b-437e-4d32-af6a-c3213841c5f1\") " pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.712003 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf669\" (UniqueName: \"kubernetes.io/projected/8512e92b-437e-4d32-af6a-c3213841c5f1-kube-api-access-lf669\") pod \"route-controller-manager-6758cd79b9-jhxvt\" (UID: \"8512e92b-437e-4d32-af6a-c3213841c5f1\") " pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.712068 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7663a79d-5498-4459-8838-61d64eeb3d6d-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.712078 4884 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7663a79d-5498-4459-8838-61d64eeb3d6d-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.712089 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7663a79d-5498-4459-8838-61d64eeb3d6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.712100 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hvdw\" (UniqueName: \"kubernetes.io/projected/7663a79d-5498-4459-8838-61d64eeb3d6d-kube-api-access-4hvdw\") on node \"crc\" DevicePath \"\"" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.712846 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8512e92b-437e-4d32-af6a-c3213841c5f1-client-ca\") pod \"route-controller-manager-6758cd79b9-jhxvt\" (UID: \"8512e92b-437e-4d32-af6a-c3213841c5f1\") " pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.713463 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8512e92b-437e-4d32-af6a-c3213841c5f1-config\") pod \"route-controller-manager-6758cd79b9-jhxvt\" (UID: \"8512e92b-437e-4d32-af6a-c3213841c5f1\") " pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.717233 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8512e92b-437e-4d32-af6a-c3213841c5f1-serving-cert\") pod \"route-controller-manager-6758cd79b9-jhxvt\" (UID: \"8512e92b-437e-4d32-af6a-c3213841c5f1\") " pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.738378 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf669\" (UniqueName: \"kubernetes.io/projected/8512e92b-437e-4d32-af6a-c3213841c5f1-kube-api-access-lf669\") pod \"route-controller-manager-6758cd79b9-jhxvt\" (UID: \"8512e92b-437e-4d32-af6a-c3213841c5f1\") " pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:50 crc kubenswrapper[4884]: I1202 01:41:50.821593 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:51 crc kubenswrapper[4884]: I1202 01:41:51.023982 4884 generic.go:334] "Generic (PLEG): container finished" podID="7663a79d-5498-4459-8838-61d64eeb3d6d" containerID="f68d77e4017e15dd19945e59720089116d7f45319bb351917853dda739ae5728" exitCode=0 Dec 02 01:41:51 crc kubenswrapper[4884]: I1202 01:41:51.024024 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" event={"ID":"7663a79d-5498-4459-8838-61d64eeb3d6d","Type":"ContainerDied","Data":"f68d77e4017e15dd19945e59720089116d7f45319bb351917853dda739ae5728"} Dec 02 01:41:51 crc kubenswrapper[4884]: I1202 01:41:51.024343 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" event={"ID":"7663a79d-5498-4459-8838-61d64eeb3d6d","Type":"ContainerDied","Data":"b064775c7c85f414d29d5f00d514b719130f5688b7e6629fe826e42062836d82"} Dec 02 01:41:51 crc kubenswrapper[4884]: I1202 01:41:51.024366 4884 scope.go:117] "RemoveContainer" containerID="f68d77e4017e15dd19945e59720089116d7f45319bb351917853dda739ae5728" Dec 02 01:41:51 crc kubenswrapper[4884]: I1202 01:41:51.024083 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx" Dec 02 01:41:51 crc kubenswrapper[4884]: I1202 01:41:51.044151 4884 scope.go:117] "RemoveContainer" containerID="f68d77e4017e15dd19945e59720089116d7f45319bb351917853dda739ae5728" Dec 02 01:41:51 crc kubenswrapper[4884]: E1202 01:41:51.044965 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f68d77e4017e15dd19945e59720089116d7f45319bb351917853dda739ae5728\": container with ID starting with f68d77e4017e15dd19945e59720089116d7f45319bb351917853dda739ae5728 not found: ID does not exist" containerID="f68d77e4017e15dd19945e59720089116d7f45319bb351917853dda739ae5728" Dec 02 01:41:51 crc kubenswrapper[4884]: I1202 01:41:51.045006 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f68d77e4017e15dd19945e59720089116d7f45319bb351917853dda739ae5728"} err="failed to get container status \"f68d77e4017e15dd19945e59720089116d7f45319bb351917853dda739ae5728\": rpc error: code = NotFound desc = could not find container \"f68d77e4017e15dd19945e59720089116d7f45319bb351917853dda739ae5728\": container with ID starting with f68d77e4017e15dd19945e59720089116d7f45319bb351917853dda739ae5728 not found: ID does not exist" Dec 02 01:41:51 crc kubenswrapper[4884]: I1202 01:41:51.057182 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx"] Dec 02 01:41:51 crc kubenswrapper[4884]: I1202 01:41:51.073321 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-577fbf7ff6-86qqx"] Dec 02 01:41:51 crc kubenswrapper[4884]: I1202 01:41:51.285434 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt"] Dec 02 01:41:51 crc kubenswrapper[4884]: W1202 01:41:51.301321 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8512e92b_437e_4d32_af6a_c3213841c5f1.slice/crio-1f2f5c92b8315ff347a430d3ac868ce4b9bea62dbe2371edfbdad51adb4d8b4d WatchSource:0}: Error finding container 1f2f5c92b8315ff347a430d3ac868ce4b9bea62dbe2371edfbdad51adb4d8b4d: Status 404 returned error can't find the container with id 1f2f5c92b8315ff347a430d3ac868ce4b9bea62dbe2371edfbdad51adb4d8b4d Dec 02 01:41:51 crc kubenswrapper[4884]: I1202 01:41:51.623407 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7663a79d-5498-4459-8838-61d64eeb3d6d" path="/var/lib/kubelet/pods/7663a79d-5498-4459-8838-61d64eeb3d6d/volumes" Dec 02 01:41:52 crc kubenswrapper[4884]: I1202 01:41:52.035259 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" event={"ID":"8512e92b-437e-4d32-af6a-c3213841c5f1","Type":"ContainerStarted","Data":"af4eb5462b8914d64b03023a92c5ff51794d26f8b17d87238ef5e249570ca5b8"} Dec 02 01:41:52 crc kubenswrapper[4884]: I1202 01:41:52.035321 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" event={"ID":"8512e92b-437e-4d32-af6a-c3213841c5f1","Type":"ContainerStarted","Data":"1f2f5c92b8315ff347a430d3ac868ce4b9bea62dbe2371edfbdad51adb4d8b4d"} Dec 02 01:41:52 crc kubenswrapper[4884]: I1202 01:41:52.035729 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:52 crc kubenswrapper[4884]: I1202 01:41:52.040857 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" Dec 02 01:41:52 crc kubenswrapper[4884]: I1202 01:41:52.061522 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6758cd79b9-jhxvt" podStartSLOduration=4.061494578 podStartE2EDuration="4.061494578s" podCreationTimestamp="2025-12-02 01:41:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:41:52.056335319 +0000 UTC m=+328.732172243" watchObservedRunningTime="2025-12-02 01:41:52.061494578 +0000 UTC m=+328.737331472" Dec 02 01:42:16 crc kubenswrapper[4884]: I1202 01:42:16.972393 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:42:16 crc kubenswrapper[4884]: I1202 01:42:16.973229 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.069386 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2mzck"] Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.072654 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2mzck" podUID="70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" containerName="registry-server" containerID="cri-o://f8c4d24df9a9b1d334b6eb38f35e5137358b21c757f512c41c4b1e3528992b88" gracePeriod=30 Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.086585 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tqtn7"] Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.086953 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tqtn7" podUID="525da77f-9387-441b-8ded-8f7b1cc41df8" containerName="registry-server" containerID="cri-o://c35594d89bfd93397dda6bc0fb6e4af987d34797d082e57b9f310ebfd560baeb" gracePeriod=30 Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.089022 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdd9f"] Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.089278 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" podUID="db61ff51-5c14-43bd-9042-cf7447ab7a3d" containerName="marketplace-operator" containerID="cri-o://e9b569ae157f96a02bdcfa134878cb182f1cbc0c5e7e423023b35ceb151dfa09" gracePeriod=30 Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.094576 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7v6v"] Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.094907 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k7v6v" podUID="25c18b0c-a129-43c2-8d8e-8381d421510a" containerName="registry-server" containerID="cri-o://76ff6d9afa54717a7d32fc7480cf0547d36d368de1a04634a492db3b9075f3f0" gracePeriod=30 Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.111073 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gcw7d"] Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.111817 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.115797 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jz4cj"] Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.116240 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jz4cj" podUID="2ef9b26a-4331-4c67-a38a-921f05f5e8c6" containerName="registry-server" containerID="cri-o://ffd58676dac6f722261f8badb316a8139566b56398b9ae2b5b76e50be89f2b0e" gracePeriod=30 Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.124238 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gcw7d"] Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.247286 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4sfk\" (UniqueName: \"kubernetes.io/projected/ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1-kube-api-access-b4sfk\") pod \"marketplace-operator-79b997595-gcw7d\" (UID: \"ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1\") " pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.247425 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gcw7d\" (UID: \"ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1\") " pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.247451 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gcw7d\" (UID: \"ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1\") " pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.304703 4884 generic.go:334] "Generic (PLEG): container finished" podID="525da77f-9387-441b-8ded-8f7b1cc41df8" containerID="c35594d89bfd93397dda6bc0fb6e4af987d34797d082e57b9f310ebfd560baeb" exitCode=0 Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.304792 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqtn7" event={"ID":"525da77f-9387-441b-8ded-8f7b1cc41df8","Type":"ContainerDied","Data":"c35594d89bfd93397dda6bc0fb6e4af987d34797d082e57b9f310ebfd560baeb"} Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.308998 4884 generic.go:334] "Generic (PLEG): container finished" podID="2ef9b26a-4331-4c67-a38a-921f05f5e8c6" containerID="ffd58676dac6f722261f8badb316a8139566b56398b9ae2b5b76e50be89f2b0e" exitCode=0 Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.309080 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jz4cj" event={"ID":"2ef9b26a-4331-4c67-a38a-921f05f5e8c6","Type":"ContainerDied","Data":"ffd58676dac6f722261f8badb316a8139566b56398b9ae2b5b76e50be89f2b0e"} Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.313521 4884 generic.go:334] "Generic (PLEG): container finished" podID="25c18b0c-a129-43c2-8d8e-8381d421510a" containerID="76ff6d9afa54717a7d32fc7480cf0547d36d368de1a04634a492db3b9075f3f0" exitCode=0 Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.313575 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7v6v" event={"ID":"25c18b0c-a129-43c2-8d8e-8381d421510a","Type":"ContainerDied","Data":"76ff6d9afa54717a7d32fc7480cf0547d36d368de1a04634a492db3b9075f3f0"} Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.315105 4884 generic.go:334] "Generic (PLEG): container finished" podID="70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" containerID="f8c4d24df9a9b1d334b6eb38f35e5137358b21c757f512c41c4b1e3528992b88" exitCode=0 Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.315149 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mzck" event={"ID":"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80","Type":"ContainerDied","Data":"f8c4d24df9a9b1d334b6eb38f35e5137358b21c757f512c41c4b1e3528992b88"} Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.316622 4884 generic.go:334] "Generic (PLEG): container finished" podID="db61ff51-5c14-43bd-9042-cf7447ab7a3d" containerID="e9b569ae157f96a02bdcfa134878cb182f1cbc0c5e7e423023b35ceb151dfa09" exitCode=0 Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.316651 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" event={"ID":"db61ff51-5c14-43bd-9042-cf7447ab7a3d","Type":"ContainerDied","Data":"e9b569ae157f96a02bdcfa134878cb182f1cbc0c5e7e423023b35ceb151dfa09"} Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.316674 4884 scope.go:117] "RemoveContainer" containerID="eb95cf030ae39b8d31a3de3fe32701ef5083b725345bb12e0d7bab83ddb770e1" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.349154 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gcw7d\" (UID: \"ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1\") " pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.349216 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gcw7d\" (UID: \"ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1\") " pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.349260 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4sfk\" (UniqueName: \"kubernetes.io/projected/ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1-kube-api-access-b4sfk\") pod \"marketplace-operator-79b997595-gcw7d\" (UID: \"ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1\") " pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.357605 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-gcw7d\" (UID: \"ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1\") " pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.357971 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-gcw7d\" (UID: \"ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1\") " pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.369834 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4sfk\" (UniqueName: \"kubernetes.io/projected/ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1-kube-api-access-b4sfk\") pod \"marketplace-operator-79b997595-gcw7d\" (UID: \"ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1\") " pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.433893 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.535969 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.559938 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.564047 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.567459 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.644483 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.656629 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-catalog-content\") pod \"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80\" (UID: \"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80\") " Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.656703 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/525da77f-9387-441b-8ded-8f7b1cc41df8-utilities\") pod \"525da77f-9387-441b-8ded-8f7b1cc41df8\" (UID: \"525da77f-9387-441b-8ded-8f7b1cc41df8\") " Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.656733 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/525da77f-9387-441b-8ded-8f7b1cc41df8-catalog-content\") pod \"525da77f-9387-441b-8ded-8f7b1cc41df8\" (UID: \"525da77f-9387-441b-8ded-8f7b1cc41df8\") " Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.656849 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-utilities\") pod \"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80\" (UID: \"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80\") " Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.656884 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pg8v8\" (UniqueName: \"kubernetes.io/projected/525da77f-9387-441b-8ded-8f7b1cc41df8-kube-api-access-pg8v8\") pod \"525da77f-9387-441b-8ded-8f7b1cc41df8\" (UID: \"525da77f-9387-441b-8ded-8f7b1cc41df8\") " Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.656905 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tx4l\" (UniqueName: \"kubernetes.io/projected/25c18b0c-a129-43c2-8d8e-8381d421510a-kube-api-access-7tx4l\") pod \"25c18b0c-a129-43c2-8d8e-8381d421510a\" (UID: \"25c18b0c-a129-43c2-8d8e-8381d421510a\") " Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.656939 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25c18b0c-a129-43c2-8d8e-8381d421510a-utilities\") pod \"25c18b0c-a129-43c2-8d8e-8381d421510a\" (UID: \"25c18b0c-a129-43c2-8d8e-8381d421510a\") " Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.656955 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25c18b0c-a129-43c2-8d8e-8381d421510a-catalog-content\") pod \"25c18b0c-a129-43c2-8d8e-8381d421510a\" (UID: \"25c18b0c-a129-43c2-8d8e-8381d421510a\") " Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.656972 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2dmz\" (UniqueName: \"kubernetes.io/projected/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-kube-api-access-m2dmz\") pod \"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80\" (UID: \"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80\") " Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.656994 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-trusted-ca\") pod \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\" (UID: \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\") " Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.657020 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsxpp\" (UniqueName: \"kubernetes.io/projected/db61ff51-5c14-43bd-9042-cf7447ab7a3d-kube-api-access-tsxpp\") pod \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\" (UID: \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\") " Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.657044 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-operator-metrics\") pod \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\" (UID: \"db61ff51-5c14-43bd-9042-cf7447ab7a3d\") " Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.658903 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25c18b0c-a129-43c2-8d8e-8381d421510a-utilities" (OuterVolumeSpecName: "utilities") pod "25c18b0c-a129-43c2-8d8e-8381d421510a" (UID: "25c18b0c-a129-43c2-8d8e-8381d421510a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.659299 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-utilities" (OuterVolumeSpecName: "utilities") pod "70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" (UID: "70d5f628-082b-4f7b-8cc9-c4a08e5d8e80"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.663137 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "db61ff51-5c14-43bd-9042-cf7447ab7a3d" (UID: "db61ff51-5c14-43bd-9042-cf7447ab7a3d"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.665707 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/525da77f-9387-441b-8ded-8f7b1cc41df8-utilities" (OuterVolumeSpecName: "utilities") pod "525da77f-9387-441b-8ded-8f7b1cc41df8" (UID: "525da77f-9387-441b-8ded-8f7b1cc41df8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.668733 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/525da77f-9387-441b-8ded-8f7b1cc41df8-kube-api-access-pg8v8" (OuterVolumeSpecName: "kube-api-access-pg8v8") pod "525da77f-9387-441b-8ded-8f7b1cc41df8" (UID: "525da77f-9387-441b-8ded-8f7b1cc41df8"). InnerVolumeSpecName "kube-api-access-pg8v8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.669289 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db61ff51-5c14-43bd-9042-cf7447ab7a3d-kube-api-access-tsxpp" (OuterVolumeSpecName: "kube-api-access-tsxpp") pod "db61ff51-5c14-43bd-9042-cf7447ab7a3d" (UID: "db61ff51-5c14-43bd-9042-cf7447ab7a3d"). InnerVolumeSpecName "kube-api-access-tsxpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.676245 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25c18b0c-a129-43c2-8d8e-8381d421510a-kube-api-access-7tx4l" (OuterVolumeSpecName: "kube-api-access-7tx4l") pod "25c18b0c-a129-43c2-8d8e-8381d421510a" (UID: "25c18b0c-a129-43c2-8d8e-8381d421510a"). InnerVolumeSpecName "kube-api-access-7tx4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.676376 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "db61ff51-5c14-43bd-9042-cf7447ab7a3d" (UID: "db61ff51-5c14-43bd-9042-cf7447ab7a3d"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.678022 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-kube-api-access-m2dmz" (OuterVolumeSpecName: "kube-api-access-m2dmz") pod "70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" (UID: "70d5f628-082b-4f7b-8cc9-c4a08e5d8e80"). InnerVolumeSpecName "kube-api-access-m2dmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.688985 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25c18b0c-a129-43c2-8d8e-8381d421510a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25c18b0c-a129-43c2-8d8e-8381d421510a" (UID: "25c18b0c-a129-43c2-8d8e-8381d421510a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.720985 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" (UID: "70d5f628-082b-4f7b-8cc9-c4a08e5d8e80"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.748884 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/525da77f-9387-441b-8ded-8f7b1cc41df8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "525da77f-9387-441b-8ded-8f7b1cc41df8" (UID: "525da77f-9387-441b-8ded-8f7b1cc41df8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.757706 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-utilities\") pod \"2ef9b26a-4331-4c67-a38a-921f05f5e8c6\" (UID: \"2ef9b26a-4331-4c67-a38a-921f05f5e8c6\") " Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.757822 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-catalog-content\") pod \"2ef9b26a-4331-4c67-a38a-921f05f5e8c6\" (UID: \"2ef9b26a-4331-4c67-a38a-921f05f5e8c6\") " Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.757871 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc8b4\" (UniqueName: \"kubernetes.io/projected/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-kube-api-access-fc8b4\") pod \"2ef9b26a-4331-4c67-a38a-921f05f5e8c6\" (UID: \"2ef9b26a-4331-4c67-a38a-921f05f5e8c6\") " Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.758148 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.758163 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pg8v8\" (UniqueName: \"kubernetes.io/projected/525da77f-9387-441b-8ded-8f7b1cc41df8-kube-api-access-pg8v8\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.758173 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tx4l\" (UniqueName: \"kubernetes.io/projected/25c18b0c-a129-43c2-8d8e-8381d421510a-kube-api-access-7tx4l\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.758182 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25c18b0c-a129-43c2-8d8e-8381d421510a-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.758192 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25c18b0c-a129-43c2-8d8e-8381d421510a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.758200 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2dmz\" (UniqueName: \"kubernetes.io/projected/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-kube-api-access-m2dmz\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.758209 4884 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.758218 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsxpp\" (UniqueName: \"kubernetes.io/projected/db61ff51-5c14-43bd-9042-cf7447ab7a3d-kube-api-access-tsxpp\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.758228 4884 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db61ff51-5c14-43bd-9042-cf7447ab7a3d-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.758243 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.758255 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/525da77f-9387-441b-8ded-8f7b1cc41df8-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.758263 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/525da77f-9387-441b-8ded-8f7b1cc41df8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.758591 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-utilities" (OuterVolumeSpecName: "utilities") pod "2ef9b26a-4331-4c67-a38a-921f05f5e8c6" (UID: "2ef9b26a-4331-4c67-a38a-921f05f5e8c6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.760562 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-kube-api-access-fc8b4" (OuterVolumeSpecName: "kube-api-access-fc8b4") pod "2ef9b26a-4331-4c67-a38a-921f05f5e8c6" (UID: "2ef9b26a-4331-4c67-a38a-921f05f5e8c6"). InnerVolumeSpecName "kube-api-access-fc8b4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.859307 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.859341 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc8b4\" (UniqueName: \"kubernetes.io/projected/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-kube-api-access-fc8b4\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.891617 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ef9b26a-4331-4c67-a38a-921f05f5e8c6" (UID: "2ef9b26a-4331-4c67-a38a-921f05f5e8c6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.897542 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-gcw7d"] Dec 02 01:42:28 crc kubenswrapper[4884]: I1202 01:42:28.960884 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ef9b26a-4331-4c67-a38a-921f05f5e8c6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.323264 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tqtn7" event={"ID":"525da77f-9387-441b-8ded-8f7b1cc41df8","Type":"ContainerDied","Data":"f879d117e5598a7c9dde048cf366ec86b4ead1ceea2cb77d8478d2b1b7c11fb6"} Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.323317 4884 scope.go:117] "RemoveContainer" containerID="c35594d89bfd93397dda6bc0fb6e4af987d34797d082e57b9f310ebfd560baeb" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.323330 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tqtn7" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.326003 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jz4cj" event={"ID":"2ef9b26a-4331-4c67-a38a-921f05f5e8c6","Type":"ContainerDied","Data":"4158e42de7e6e5c77c26d34d00d269e648fa03cf0b2357d075442d3823541eec"} Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.326084 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jz4cj" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.341353 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7v6v" event={"ID":"25c18b0c-a129-43c2-8d8e-8381d421510a","Type":"ContainerDied","Data":"1ffda3e310519bc3f003497cefb591b20a42ab90c9adbd89cb1abf4a3037ac2b"} Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.341402 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7v6v" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.345783 4884 scope.go:117] "RemoveContainer" containerID="1b962ae3c89230dbba89742da2888e20ca6e6d0f08a8a4b868ce6b167108e7c5" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.348193 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" event={"ID":"ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1","Type":"ContainerStarted","Data":"e7f5273f8f70e0d6ae00d9cbf70fc9f022dc4d102f199feb8a7a9a2b11409639"} Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.348240 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" event={"ID":"ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1","Type":"ContainerStarted","Data":"97db144f64f9f14d03ada0d2f12fb38350a61de903404f933f60621c633d68f0"} Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.348966 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.350357 4884 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-gcw7d container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" start-of-body= Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.350409 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" podUID="ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.62:8080/healthz\": dial tcp 10.217.0.62:8080: connect: connection refused" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.352786 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mzck" event={"ID":"70d5f628-082b-4f7b-8cc9-c4a08e5d8e80","Type":"ContainerDied","Data":"050bbf1f3c94858339b102e0a95d2d4bed2f6bd7429cde8dd796a406c8f59a07"} Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.352868 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2mzck" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.354350 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" event={"ID":"db61ff51-5c14-43bd-9042-cf7447ab7a3d","Type":"ContainerDied","Data":"f6711810588bb501331bd68ac442945f4645aa34538d98f345484b9bfdddf167"} Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.354513 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rdd9f" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.364990 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tqtn7"] Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.367757 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tqtn7"] Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.391519 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" podStartSLOduration=1.39150117 podStartE2EDuration="1.39150117s" podCreationTimestamp="2025-12-02 01:42:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:42:29.3775525 +0000 UTC m=+366.053389394" watchObservedRunningTime="2025-12-02 01:42:29.39150117 +0000 UTC m=+366.067338054" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.396956 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7v6v"] Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.401139 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7v6v"] Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.409241 4884 scope.go:117] "RemoveContainer" containerID="5c99f3ee4ed330750d81af7ac344010f66fc61f741191b7320a2a5de40434169" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.410915 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jz4cj"] Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.420048 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jz4cj"] Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.425328 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2mzck"] Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.426307 4884 scope.go:117] "RemoveContainer" containerID="ffd58676dac6f722261f8badb316a8139566b56398b9ae2b5b76e50be89f2b0e" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.436730 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2mzck"] Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.445144 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdd9f"] Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.449475 4884 scope.go:117] "RemoveContainer" containerID="29b31ec9e2fcb4ef073b210ded678e61ed3288333b6a4f8a6728257960bb3781" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.451616 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdd9f"] Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.466088 4884 scope.go:117] "RemoveContainer" containerID="4887c8c8c8c34c575d931495a2a8cb256102927aae2e3c93010b5af1fd724ef8" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.483219 4884 scope.go:117] "RemoveContainer" containerID="76ff6d9afa54717a7d32fc7480cf0547d36d368de1a04634a492db3b9075f3f0" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.503112 4884 scope.go:117] "RemoveContainer" containerID="92221f554ae5054dc1ef6472f2f4cad275510a4e5186db81750a20598d4f2ea6" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.518826 4884 scope.go:117] "RemoveContainer" containerID="7c60316071285a21c46e7bcceff9a79b80407dfe0777a0f26216ab7c8d45eb30" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.533285 4884 scope.go:117] "RemoveContainer" containerID="f8c4d24df9a9b1d334b6eb38f35e5137358b21c757f512c41c4b1e3528992b88" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.549837 4884 scope.go:117] "RemoveContainer" containerID="1e671dbaf37ae0d9e526c73efb5e65d452d25ae29aa8028e8617bae03cbd397e" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.564371 4884 scope.go:117] "RemoveContainer" containerID="a1774190940f7913f4531396bd52b132b0f1f6578b889bfed8aa7b93bc56aa12" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.580906 4884 scope.go:117] "RemoveContainer" containerID="e9b569ae157f96a02bdcfa134878cb182f1cbc0c5e7e423023b35ceb151dfa09" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.639001 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25c18b0c-a129-43c2-8d8e-8381d421510a" path="/var/lib/kubelet/pods/25c18b0c-a129-43c2-8d8e-8381d421510a/volumes" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.639792 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ef9b26a-4331-4c67-a38a-921f05f5e8c6" path="/var/lib/kubelet/pods/2ef9b26a-4331-4c67-a38a-921f05f5e8c6/volumes" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.640426 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="525da77f-9387-441b-8ded-8f7b1cc41df8" path="/var/lib/kubelet/pods/525da77f-9387-441b-8ded-8f7b1cc41df8/volumes" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.641513 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" path="/var/lib/kubelet/pods/70d5f628-082b-4f7b-8cc9-c4a08e5d8e80/volumes" Dec 02 01:42:29 crc kubenswrapper[4884]: I1202 01:42:29.642325 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db61ff51-5c14-43bd-9042-cf7447ab7a3d" path="/var/lib/kubelet/pods/db61ff51-5c14-43bd-9042-cf7447ab7a3d/volumes" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.289863 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dddxc"] Dec 02 01:42:30 crc kubenswrapper[4884]: E1202 01:42:30.290457 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" containerName="extract-content" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.290600 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" containerName="extract-content" Dec 02 01:42:30 crc kubenswrapper[4884]: E1202 01:42:30.290725 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25c18b0c-a129-43c2-8d8e-8381d421510a" containerName="extract-utilities" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.290948 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="25c18b0c-a129-43c2-8d8e-8381d421510a" containerName="extract-utilities" Dec 02 01:42:30 crc kubenswrapper[4884]: E1202 01:42:30.291081 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25c18b0c-a129-43c2-8d8e-8381d421510a" containerName="registry-server" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.291194 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="25c18b0c-a129-43c2-8d8e-8381d421510a" containerName="registry-server" Dec 02 01:42:30 crc kubenswrapper[4884]: E1202 01:42:30.291322 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="525da77f-9387-441b-8ded-8f7b1cc41df8" containerName="registry-server" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.291433 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="525da77f-9387-441b-8ded-8f7b1cc41df8" containerName="registry-server" Dec 02 01:42:30 crc kubenswrapper[4884]: E1202 01:42:30.291546 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db61ff51-5c14-43bd-9042-cf7447ab7a3d" containerName="marketplace-operator" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.291662 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="db61ff51-5c14-43bd-9042-cf7447ab7a3d" containerName="marketplace-operator" Dec 02 01:42:30 crc kubenswrapper[4884]: E1202 01:42:30.291806 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ef9b26a-4331-4c67-a38a-921f05f5e8c6" containerName="extract-utilities" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.291919 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ef9b26a-4331-4c67-a38a-921f05f5e8c6" containerName="extract-utilities" Dec 02 01:42:30 crc kubenswrapper[4884]: E1202 01:42:30.292048 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="525da77f-9387-441b-8ded-8f7b1cc41df8" containerName="extract-content" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.292210 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="525da77f-9387-441b-8ded-8f7b1cc41df8" containerName="extract-content" Dec 02 01:42:30 crc kubenswrapper[4884]: E1202 01:42:30.292328 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ef9b26a-4331-4c67-a38a-921f05f5e8c6" containerName="extract-content" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.292433 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ef9b26a-4331-4c67-a38a-921f05f5e8c6" containerName="extract-content" Dec 02 01:42:30 crc kubenswrapper[4884]: E1202 01:42:30.292543 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" containerName="extract-utilities" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.292662 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" containerName="extract-utilities" Dec 02 01:42:30 crc kubenswrapper[4884]: E1202 01:42:30.292808 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ef9b26a-4331-4c67-a38a-921f05f5e8c6" containerName="registry-server" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.292943 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ef9b26a-4331-4c67-a38a-921f05f5e8c6" containerName="registry-server" Dec 02 01:42:30 crc kubenswrapper[4884]: E1202 01:42:30.293066 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="525da77f-9387-441b-8ded-8f7b1cc41df8" containerName="extract-utilities" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.293175 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="525da77f-9387-441b-8ded-8f7b1cc41df8" containerName="extract-utilities" Dec 02 01:42:30 crc kubenswrapper[4884]: E1202 01:42:30.293286 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" containerName="registry-server" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.293390 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" containerName="registry-server" Dec 02 01:42:30 crc kubenswrapper[4884]: E1202 01:42:30.293504 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25c18b0c-a129-43c2-8d8e-8381d421510a" containerName="extract-content" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.293629 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="25c18b0c-a129-43c2-8d8e-8381d421510a" containerName="extract-content" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.293937 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="70d5f628-082b-4f7b-8cc9-c4a08e5d8e80" containerName="registry-server" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.294069 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="25c18b0c-a129-43c2-8d8e-8381d421510a" containerName="registry-server" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.294180 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ef9b26a-4331-4c67-a38a-921f05f5e8c6" containerName="registry-server" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.294294 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="525da77f-9387-441b-8ded-8f7b1cc41df8" containerName="registry-server" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.294410 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="db61ff51-5c14-43bd-9042-cf7447ab7a3d" containerName="marketplace-operator" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.294546 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="db61ff51-5c14-43bd-9042-cf7447ab7a3d" containerName="marketplace-operator" Dec 02 01:42:30 crc kubenswrapper[4884]: E1202 01:42:30.294886 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db61ff51-5c14-43bd-9042-cf7447ab7a3d" containerName="marketplace-operator" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.295029 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="db61ff51-5c14-43bd-9042-cf7447ab7a3d" containerName="marketplace-operator" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.296535 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dddxc" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.299018 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.301784 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dddxc"] Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.370820 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-gcw7d" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.376866 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqrlz\" (UniqueName: \"kubernetes.io/projected/cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85-kube-api-access-nqrlz\") pod \"redhat-marketplace-dddxc\" (UID: \"cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85\") " pod="openshift-marketplace/redhat-marketplace-dddxc" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.376942 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85-utilities\") pod \"redhat-marketplace-dddxc\" (UID: \"cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85\") " pod="openshift-marketplace/redhat-marketplace-dddxc" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.376969 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85-catalog-content\") pod \"redhat-marketplace-dddxc\" (UID: \"cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85\") " pod="openshift-marketplace/redhat-marketplace-dddxc" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.478016 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqrlz\" (UniqueName: \"kubernetes.io/projected/cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85-kube-api-access-nqrlz\") pod \"redhat-marketplace-dddxc\" (UID: \"cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85\") " pod="openshift-marketplace/redhat-marketplace-dddxc" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.478292 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85-utilities\") pod \"redhat-marketplace-dddxc\" (UID: \"cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85\") " pod="openshift-marketplace/redhat-marketplace-dddxc" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.478421 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85-catalog-content\") pod \"redhat-marketplace-dddxc\" (UID: \"cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85\") " pod="openshift-marketplace/redhat-marketplace-dddxc" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.479938 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85-utilities\") pod \"redhat-marketplace-dddxc\" (UID: \"cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85\") " pod="openshift-marketplace/redhat-marketplace-dddxc" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.480332 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85-catalog-content\") pod \"redhat-marketplace-dddxc\" (UID: \"cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85\") " pod="openshift-marketplace/redhat-marketplace-dddxc" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.486286 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bmrjj"] Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.488042 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bmrjj" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.492419 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bmrjj"] Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.494689 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.507638 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqrlz\" (UniqueName: \"kubernetes.io/projected/cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85-kube-api-access-nqrlz\") pod \"redhat-marketplace-dddxc\" (UID: \"cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85\") " pod="openshift-marketplace/redhat-marketplace-dddxc" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.580304 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81df2b19-2e90-4a01-b333-b4b072229d90-utilities\") pod \"redhat-operators-bmrjj\" (UID: \"81df2b19-2e90-4a01-b333-b4b072229d90\") " pod="openshift-marketplace/redhat-operators-bmrjj" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.580560 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnfjw\" (UniqueName: \"kubernetes.io/projected/81df2b19-2e90-4a01-b333-b4b072229d90-kube-api-access-nnfjw\") pod \"redhat-operators-bmrjj\" (UID: \"81df2b19-2e90-4a01-b333-b4b072229d90\") " pod="openshift-marketplace/redhat-operators-bmrjj" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.581111 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81df2b19-2e90-4a01-b333-b4b072229d90-catalog-content\") pod \"redhat-operators-bmrjj\" (UID: \"81df2b19-2e90-4a01-b333-b4b072229d90\") " pod="openshift-marketplace/redhat-operators-bmrjj" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.613265 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dddxc" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.687635 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnfjw\" (UniqueName: \"kubernetes.io/projected/81df2b19-2e90-4a01-b333-b4b072229d90-kube-api-access-nnfjw\") pod \"redhat-operators-bmrjj\" (UID: \"81df2b19-2e90-4a01-b333-b4b072229d90\") " pod="openshift-marketplace/redhat-operators-bmrjj" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.688407 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81df2b19-2e90-4a01-b333-b4b072229d90-catalog-content\") pod \"redhat-operators-bmrjj\" (UID: \"81df2b19-2e90-4a01-b333-b4b072229d90\") " pod="openshift-marketplace/redhat-operators-bmrjj" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.688622 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81df2b19-2e90-4a01-b333-b4b072229d90-utilities\") pod \"redhat-operators-bmrjj\" (UID: \"81df2b19-2e90-4a01-b333-b4b072229d90\") " pod="openshift-marketplace/redhat-operators-bmrjj" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.691543 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81df2b19-2e90-4a01-b333-b4b072229d90-catalog-content\") pod \"redhat-operators-bmrjj\" (UID: \"81df2b19-2e90-4a01-b333-b4b072229d90\") " pod="openshift-marketplace/redhat-operators-bmrjj" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.695805 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81df2b19-2e90-4a01-b333-b4b072229d90-utilities\") pod \"redhat-operators-bmrjj\" (UID: \"81df2b19-2e90-4a01-b333-b4b072229d90\") " pod="openshift-marketplace/redhat-operators-bmrjj" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.722641 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnfjw\" (UniqueName: \"kubernetes.io/projected/81df2b19-2e90-4a01-b333-b4b072229d90-kube-api-access-nnfjw\") pod \"redhat-operators-bmrjj\" (UID: \"81df2b19-2e90-4a01-b333-b4b072229d90\") " pod="openshift-marketplace/redhat-operators-bmrjj" Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.808530 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bmrjj" Dec 02 01:42:30 crc kubenswrapper[4884]: W1202 01:42:30.828096 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd3c5eb9_e3ac_431f_88ee_916b5d9a7b85.slice/crio-d3ece75ef4b28e41e9665c4a064daf50db74529033a0f9c8a0bc56d1d539739c WatchSource:0}: Error finding container d3ece75ef4b28e41e9665c4a064daf50db74529033a0f9c8a0bc56d1d539739c: Status 404 returned error can't find the container with id d3ece75ef4b28e41e9665c4a064daf50db74529033a0f9c8a0bc56d1d539739c Dec 02 01:42:30 crc kubenswrapper[4884]: I1202 01:42:30.831415 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dddxc"] Dec 02 01:42:31 crc kubenswrapper[4884]: I1202 01:42:31.272012 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bmrjj"] Dec 02 01:42:31 crc kubenswrapper[4884]: W1202 01:42:31.274352 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81df2b19_2e90_4a01_b333_b4b072229d90.slice/crio-1fc9c34644e58dc877aba5c3b485721e31191f716987839192cfb147e890414e WatchSource:0}: Error finding container 1fc9c34644e58dc877aba5c3b485721e31191f716987839192cfb147e890414e: Status 404 returned error can't find the container with id 1fc9c34644e58dc877aba5c3b485721e31191f716987839192cfb147e890414e Dec 02 01:42:31 crc kubenswrapper[4884]: I1202 01:42:31.384038 4884 generic.go:334] "Generic (PLEG): container finished" podID="cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85" containerID="fb0f267edebdd783a8455f4c866b122a470d7a327b98f19e65d930c051693860" exitCode=0 Dec 02 01:42:31 crc kubenswrapper[4884]: I1202 01:42:31.384229 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dddxc" event={"ID":"cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85","Type":"ContainerDied","Data":"fb0f267edebdd783a8455f4c866b122a470d7a327b98f19e65d930c051693860"} Dec 02 01:42:31 crc kubenswrapper[4884]: I1202 01:42:31.384310 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dddxc" event={"ID":"cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85","Type":"ContainerStarted","Data":"d3ece75ef4b28e41e9665c4a064daf50db74529033a0f9c8a0bc56d1d539739c"} Dec 02 01:42:31 crc kubenswrapper[4884]: I1202 01:42:31.385778 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmrjj" event={"ID":"81df2b19-2e90-4a01-b333-b4b072229d90","Type":"ContainerStarted","Data":"1fc9c34644e58dc877aba5c3b485721e31191f716987839192cfb147e890414e"} Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.395486 4884 generic.go:334] "Generic (PLEG): container finished" podID="cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85" containerID="9eb1152b14085ad901a273b23f088c2da8ef088d36cd67d3e1698e9433ee63b8" exitCode=0 Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.395571 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dddxc" event={"ID":"cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85","Type":"ContainerDied","Data":"9eb1152b14085ad901a273b23f088c2da8ef088d36cd67d3e1698e9433ee63b8"} Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.397550 4884 generic.go:334] "Generic (PLEG): container finished" podID="81df2b19-2e90-4a01-b333-b4b072229d90" containerID="54da48ed5f3d3cfa85c6ff7c2c0a5a08c4d78edcca7212ec0daf16e7dfdeefe8" exitCode=0 Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.398523 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmrjj" event={"ID":"81df2b19-2e90-4a01-b333-b4b072229d90","Type":"ContainerDied","Data":"54da48ed5f3d3cfa85c6ff7c2c0a5a08c4d78edcca7212ec0daf16e7dfdeefe8"} Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.686487 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jspph"] Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.687833 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jspph" Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.689926 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.756256 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jspph"] Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.837240 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80d84c1-0c7c-44ff-b7f2-fe4704103241-catalog-content\") pod \"certified-operators-jspph\" (UID: \"a80d84c1-0c7c-44ff-b7f2-fe4704103241\") " pod="openshift-marketplace/certified-operators-jspph" Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.837471 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80d84c1-0c7c-44ff-b7f2-fe4704103241-utilities\") pod \"certified-operators-jspph\" (UID: \"a80d84c1-0c7c-44ff-b7f2-fe4704103241\") " pod="openshift-marketplace/certified-operators-jspph" Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.837543 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhjhm\" (UniqueName: \"kubernetes.io/projected/a80d84c1-0c7c-44ff-b7f2-fe4704103241-kube-api-access-hhjhm\") pod \"certified-operators-jspph\" (UID: \"a80d84c1-0c7c-44ff-b7f2-fe4704103241\") " pod="openshift-marketplace/certified-operators-jspph" Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.886154 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p4sd7"] Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.887363 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4sd7" Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.889610 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.893071 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4sd7"] Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.938641 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhjhm\" (UniqueName: \"kubernetes.io/projected/a80d84c1-0c7c-44ff-b7f2-fe4704103241-kube-api-access-hhjhm\") pod \"certified-operators-jspph\" (UID: \"a80d84c1-0c7c-44ff-b7f2-fe4704103241\") " pod="openshift-marketplace/certified-operators-jspph" Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.938709 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80d84c1-0c7c-44ff-b7f2-fe4704103241-catalog-content\") pod \"certified-operators-jspph\" (UID: \"a80d84c1-0c7c-44ff-b7f2-fe4704103241\") " pod="openshift-marketplace/certified-operators-jspph" Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.938782 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80d84c1-0c7c-44ff-b7f2-fe4704103241-utilities\") pod \"certified-operators-jspph\" (UID: \"a80d84c1-0c7c-44ff-b7f2-fe4704103241\") " pod="openshift-marketplace/certified-operators-jspph" Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.939244 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a80d84c1-0c7c-44ff-b7f2-fe4704103241-catalog-content\") pod \"certified-operators-jspph\" (UID: \"a80d84c1-0c7c-44ff-b7f2-fe4704103241\") " pod="openshift-marketplace/certified-operators-jspph" Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.940654 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a80d84c1-0c7c-44ff-b7f2-fe4704103241-utilities\") pod \"certified-operators-jspph\" (UID: \"a80d84c1-0c7c-44ff-b7f2-fe4704103241\") " pod="openshift-marketplace/certified-operators-jspph" Dec 02 01:42:32 crc kubenswrapper[4884]: I1202 01:42:32.958062 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhjhm\" (UniqueName: \"kubernetes.io/projected/a80d84c1-0c7c-44ff-b7f2-fe4704103241-kube-api-access-hhjhm\") pod \"certified-operators-jspph\" (UID: \"a80d84c1-0c7c-44ff-b7f2-fe4704103241\") " pod="openshift-marketplace/certified-operators-jspph" Dec 02 01:42:33 crc kubenswrapper[4884]: I1202 01:42:33.039491 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/154cc736-2f3e-4254-835f-c32fe8557341-catalog-content\") pod \"community-operators-p4sd7\" (UID: \"154cc736-2f3e-4254-835f-c32fe8557341\") " pod="openshift-marketplace/community-operators-p4sd7" Dec 02 01:42:33 crc kubenswrapper[4884]: I1202 01:42:33.039604 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5tqd\" (UniqueName: \"kubernetes.io/projected/154cc736-2f3e-4254-835f-c32fe8557341-kube-api-access-d5tqd\") pod \"community-operators-p4sd7\" (UID: \"154cc736-2f3e-4254-835f-c32fe8557341\") " pod="openshift-marketplace/community-operators-p4sd7" Dec 02 01:42:33 crc kubenswrapper[4884]: I1202 01:42:33.039653 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/154cc736-2f3e-4254-835f-c32fe8557341-utilities\") pod \"community-operators-p4sd7\" (UID: \"154cc736-2f3e-4254-835f-c32fe8557341\") " pod="openshift-marketplace/community-operators-p4sd7" Dec 02 01:42:33 crc kubenswrapper[4884]: I1202 01:42:33.059905 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jspph" Dec 02 01:42:33 crc kubenswrapper[4884]: I1202 01:42:33.140644 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/154cc736-2f3e-4254-835f-c32fe8557341-utilities\") pod \"community-operators-p4sd7\" (UID: \"154cc736-2f3e-4254-835f-c32fe8557341\") " pod="openshift-marketplace/community-operators-p4sd7" Dec 02 01:42:33 crc kubenswrapper[4884]: I1202 01:42:33.141034 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/154cc736-2f3e-4254-835f-c32fe8557341-catalog-content\") pod \"community-operators-p4sd7\" (UID: \"154cc736-2f3e-4254-835f-c32fe8557341\") " pod="openshift-marketplace/community-operators-p4sd7" Dec 02 01:42:33 crc kubenswrapper[4884]: I1202 01:42:33.141076 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5tqd\" (UniqueName: \"kubernetes.io/projected/154cc736-2f3e-4254-835f-c32fe8557341-kube-api-access-d5tqd\") pod \"community-operators-p4sd7\" (UID: \"154cc736-2f3e-4254-835f-c32fe8557341\") " pod="openshift-marketplace/community-operators-p4sd7" Dec 02 01:42:33 crc kubenswrapper[4884]: I1202 01:42:33.142255 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/154cc736-2f3e-4254-835f-c32fe8557341-utilities\") pod \"community-operators-p4sd7\" (UID: \"154cc736-2f3e-4254-835f-c32fe8557341\") " pod="openshift-marketplace/community-operators-p4sd7" Dec 02 01:42:33 crc kubenswrapper[4884]: I1202 01:42:33.142517 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/154cc736-2f3e-4254-835f-c32fe8557341-catalog-content\") pod \"community-operators-p4sd7\" (UID: \"154cc736-2f3e-4254-835f-c32fe8557341\") " pod="openshift-marketplace/community-operators-p4sd7" Dec 02 01:42:33 crc kubenswrapper[4884]: I1202 01:42:33.160993 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5tqd\" (UniqueName: \"kubernetes.io/projected/154cc736-2f3e-4254-835f-c32fe8557341-kube-api-access-d5tqd\") pod \"community-operators-p4sd7\" (UID: \"154cc736-2f3e-4254-835f-c32fe8557341\") " pod="openshift-marketplace/community-operators-p4sd7" Dec 02 01:42:33 crc kubenswrapper[4884]: I1202 01:42:33.209693 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4sd7" Dec 02 01:42:33 crc kubenswrapper[4884]: I1202 01:42:33.450215 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jspph"] Dec 02 01:42:33 crc kubenswrapper[4884]: W1202 01:42:33.468869 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda80d84c1_0c7c_44ff_b7f2_fe4704103241.slice/crio-bce0db55154c5e0c5b1f0513d735a067a41facc1a514f722d90538f32290c1f1 WatchSource:0}: Error finding container bce0db55154c5e0c5b1f0513d735a067a41facc1a514f722d90538f32290c1f1: Status 404 returned error can't find the container with id bce0db55154c5e0c5b1f0513d735a067a41facc1a514f722d90538f32290c1f1 Dec 02 01:42:33 crc kubenswrapper[4884]: I1202 01:42:33.599317 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4sd7"] Dec 02 01:42:33 crc kubenswrapper[4884]: W1202 01:42:33.605299 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod154cc736_2f3e_4254_835f_c32fe8557341.slice/crio-62c1728d7d0b65edaa447d2b80d31770006b6c8e6f49246b7fe1cac3c0dc6e47 WatchSource:0}: Error finding container 62c1728d7d0b65edaa447d2b80d31770006b6c8e6f49246b7fe1cac3c0dc6e47: Status 404 returned error can't find the container with id 62c1728d7d0b65edaa447d2b80d31770006b6c8e6f49246b7fe1cac3c0dc6e47 Dec 02 01:42:34 crc kubenswrapper[4884]: I1202 01:42:34.411101 4884 generic.go:334] "Generic (PLEG): container finished" podID="154cc736-2f3e-4254-835f-c32fe8557341" containerID="b4f8e1d99e48fc53c1aa6552ce36e93745c75be840173668a672990f8cadb7bb" exitCode=0 Dec 02 01:42:34 crc kubenswrapper[4884]: I1202 01:42:34.411169 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4sd7" event={"ID":"154cc736-2f3e-4254-835f-c32fe8557341","Type":"ContainerDied","Data":"b4f8e1d99e48fc53c1aa6552ce36e93745c75be840173668a672990f8cadb7bb"} Dec 02 01:42:34 crc kubenswrapper[4884]: I1202 01:42:34.411547 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4sd7" event={"ID":"154cc736-2f3e-4254-835f-c32fe8557341","Type":"ContainerStarted","Data":"62c1728d7d0b65edaa447d2b80d31770006b6c8e6f49246b7fe1cac3c0dc6e47"} Dec 02 01:42:34 crc kubenswrapper[4884]: I1202 01:42:34.413881 4884 generic.go:334] "Generic (PLEG): container finished" podID="81df2b19-2e90-4a01-b333-b4b072229d90" containerID="efd90844dac0b977574189e15d687e20d5a92f84c5e3ef7e2554a6a0dcea0e99" exitCode=0 Dec 02 01:42:34 crc kubenswrapper[4884]: I1202 01:42:34.413931 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmrjj" event={"ID":"81df2b19-2e90-4a01-b333-b4b072229d90","Type":"ContainerDied","Data":"efd90844dac0b977574189e15d687e20d5a92f84c5e3ef7e2554a6a0dcea0e99"} Dec 02 01:42:34 crc kubenswrapper[4884]: I1202 01:42:34.416964 4884 generic.go:334] "Generic (PLEG): container finished" podID="a80d84c1-0c7c-44ff-b7f2-fe4704103241" containerID="f7b84d11ed82471bdce0f16b06535aa5e1f10ef35cd8169f550950c27eeabc81" exitCode=0 Dec 02 01:42:34 crc kubenswrapper[4884]: I1202 01:42:34.417012 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jspph" event={"ID":"a80d84c1-0c7c-44ff-b7f2-fe4704103241","Type":"ContainerDied","Data":"f7b84d11ed82471bdce0f16b06535aa5e1f10ef35cd8169f550950c27eeabc81"} Dec 02 01:42:34 crc kubenswrapper[4884]: I1202 01:42:34.417028 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jspph" event={"ID":"a80d84c1-0c7c-44ff-b7f2-fe4704103241","Type":"ContainerStarted","Data":"bce0db55154c5e0c5b1f0513d735a067a41facc1a514f722d90538f32290c1f1"} Dec 02 01:42:34 crc kubenswrapper[4884]: I1202 01:42:34.422952 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dddxc" event={"ID":"cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85","Type":"ContainerStarted","Data":"c1cc10628036129d7da5fb62fcd8bef2163aaa1bcbecc1893243ab4afca3c333"} Dec 02 01:42:34 crc kubenswrapper[4884]: I1202 01:42:34.458770 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dddxc" podStartSLOduration=2.502095364 podStartE2EDuration="4.458731181s" podCreationTimestamp="2025-12-02 01:42:30 +0000 UTC" firstStartedPulling="2025-12-02 01:42:31.388805819 +0000 UTC m=+368.064642703" lastFinishedPulling="2025-12-02 01:42:33.345441626 +0000 UTC m=+370.021278520" observedRunningTime="2025-12-02 01:42:34.453202842 +0000 UTC m=+371.129039726" watchObservedRunningTime="2025-12-02 01:42:34.458731181 +0000 UTC m=+371.134568065" Dec 02 01:42:35 crc kubenswrapper[4884]: I1202 01:42:35.430913 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bmrjj" event={"ID":"81df2b19-2e90-4a01-b333-b4b072229d90","Type":"ContainerStarted","Data":"6d9c87d32852e4e2dc6fc6bf215b3d62ec8f6bec4db58a4df9a36bbebcab8602"} Dec 02 01:42:35 crc kubenswrapper[4884]: I1202 01:42:35.433085 4884 generic.go:334] "Generic (PLEG): container finished" podID="a80d84c1-0c7c-44ff-b7f2-fe4704103241" containerID="699278d98051fbb902f149726d80582a306ed6aa5f5e1b32c3db3fccba01275f" exitCode=0 Dec 02 01:42:35 crc kubenswrapper[4884]: I1202 01:42:35.433149 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jspph" event={"ID":"a80d84c1-0c7c-44ff-b7f2-fe4704103241","Type":"ContainerDied","Data":"699278d98051fbb902f149726d80582a306ed6aa5f5e1b32c3db3fccba01275f"} Dec 02 01:42:35 crc kubenswrapper[4884]: I1202 01:42:35.437639 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4sd7" event={"ID":"154cc736-2f3e-4254-835f-c32fe8557341","Type":"ContainerStarted","Data":"5e21f165e7439b273620c0cd433721b515c7c70a86b63a1deebf9360ac48a4a3"} Dec 02 01:42:35 crc kubenswrapper[4884]: I1202 01:42:35.455446 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bmrjj" podStartSLOduration=3.005399488 podStartE2EDuration="5.455418441s" podCreationTimestamp="2025-12-02 01:42:30 +0000 UTC" firstStartedPulling="2025-12-02 01:42:32.400100025 +0000 UTC m=+369.075936919" lastFinishedPulling="2025-12-02 01:42:34.850118988 +0000 UTC m=+371.525955872" observedRunningTime="2025-12-02 01:42:35.448969379 +0000 UTC m=+372.124806263" watchObservedRunningTime="2025-12-02 01:42:35.455418441 +0000 UTC m=+372.131255325" Dec 02 01:42:36 crc kubenswrapper[4884]: I1202 01:42:36.444212 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jspph" event={"ID":"a80d84c1-0c7c-44ff-b7f2-fe4704103241","Type":"ContainerStarted","Data":"711f203cffa58b153450abff15703cec67a4669b7c3bcaccc072118adcef5293"} Dec 02 01:42:36 crc kubenswrapper[4884]: I1202 01:42:36.446292 4884 generic.go:334] "Generic (PLEG): container finished" podID="154cc736-2f3e-4254-835f-c32fe8557341" containerID="5e21f165e7439b273620c0cd433721b515c7c70a86b63a1deebf9360ac48a4a3" exitCode=0 Dec 02 01:42:36 crc kubenswrapper[4884]: I1202 01:42:36.446385 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4sd7" event={"ID":"154cc736-2f3e-4254-835f-c32fe8557341","Type":"ContainerDied","Data":"5e21f165e7439b273620c0cd433721b515c7c70a86b63a1deebf9360ac48a4a3"} Dec 02 01:42:36 crc kubenswrapper[4884]: I1202 01:42:36.461761 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jspph" podStartSLOduration=2.990695313 podStartE2EDuration="4.461728341s" podCreationTimestamp="2025-12-02 01:42:32 +0000 UTC" firstStartedPulling="2025-12-02 01:42:34.41845092 +0000 UTC m=+371.094287804" lastFinishedPulling="2025-12-02 01:42:35.889483938 +0000 UTC m=+372.565320832" observedRunningTime="2025-12-02 01:42:36.459454884 +0000 UTC m=+373.135291778" watchObservedRunningTime="2025-12-02 01:42:36.461728341 +0000 UTC m=+373.137565225" Dec 02 01:42:38 crc kubenswrapper[4884]: I1202 01:42:38.458441 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4sd7" event={"ID":"154cc736-2f3e-4254-835f-c32fe8557341","Type":"ContainerStarted","Data":"d8401805d7d52edf5a1438252ac98714229c74512ed740908268e10cc8ead12f"} Dec 02 01:42:40 crc kubenswrapper[4884]: I1202 01:42:40.613826 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dddxc" Dec 02 01:42:40 crc kubenswrapper[4884]: I1202 01:42:40.614409 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dddxc" Dec 02 01:42:40 crc kubenswrapper[4884]: I1202 01:42:40.656266 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dddxc" Dec 02 01:42:40 crc kubenswrapper[4884]: I1202 01:42:40.683208 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p4sd7" podStartSLOduration=6.143848025 podStartE2EDuration="8.683177718s" podCreationTimestamp="2025-12-02 01:42:32 +0000 UTC" firstStartedPulling="2025-12-02 01:42:34.413171848 +0000 UTC m=+371.089008742" lastFinishedPulling="2025-12-02 01:42:36.952501551 +0000 UTC m=+373.628338435" observedRunningTime="2025-12-02 01:42:38.481028641 +0000 UTC m=+375.156865515" watchObservedRunningTime="2025-12-02 01:42:40.683177718 +0000 UTC m=+377.359014612" Dec 02 01:42:40 crc kubenswrapper[4884]: I1202 01:42:40.809533 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bmrjj" Dec 02 01:42:40 crc kubenswrapper[4884]: I1202 01:42:40.809623 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bmrjj" Dec 02 01:42:40 crc kubenswrapper[4884]: I1202 01:42:40.855366 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bmrjj" Dec 02 01:42:41 crc kubenswrapper[4884]: I1202 01:42:41.533296 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dddxc" Dec 02 01:42:41 crc kubenswrapper[4884]: I1202 01:42:41.536894 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bmrjj" Dec 02 01:42:43 crc kubenswrapper[4884]: I1202 01:42:43.060422 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jspph" Dec 02 01:42:43 crc kubenswrapper[4884]: I1202 01:42:43.061088 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jspph" Dec 02 01:42:43 crc kubenswrapper[4884]: I1202 01:42:43.102999 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jspph" Dec 02 01:42:43 crc kubenswrapper[4884]: I1202 01:42:43.210583 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p4sd7" Dec 02 01:42:43 crc kubenswrapper[4884]: I1202 01:42:43.210669 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p4sd7" Dec 02 01:42:43 crc kubenswrapper[4884]: I1202 01:42:43.252660 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p4sd7" Dec 02 01:42:43 crc kubenswrapper[4884]: I1202 01:42:43.527626 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jspph" Dec 02 01:42:43 crc kubenswrapper[4884]: I1202 01:42:43.530846 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p4sd7" Dec 02 01:42:46 crc kubenswrapper[4884]: I1202 01:42:46.971565 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:42:46 crc kubenswrapper[4884]: I1202 01:42:46.971904 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:42:48 crc kubenswrapper[4884]: I1202 01:42:48.473980 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-66f89d6466-dr6vt"] Dec 02 01:42:48 crc kubenswrapper[4884]: I1202 01:42:48.474227 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" podUID="232b9fb4-1281-4f12-ab4e-8b00aa0d4002" containerName="controller-manager" containerID="cri-o://b38d6faf09251e97809f01685c0c41d439d963eb6868838a8b0d8de4aa86193f" gracePeriod=30 Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.524693 4884 generic.go:334] "Generic (PLEG): container finished" podID="232b9fb4-1281-4f12-ab4e-8b00aa0d4002" containerID="b38d6faf09251e97809f01685c0c41d439d963eb6868838a8b0d8de4aa86193f" exitCode=0 Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.524950 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" event={"ID":"232b9fb4-1281-4f12-ab4e-8b00aa0d4002","Type":"ContainerDied","Data":"b38d6faf09251e97809f01685c0c41d439d963eb6868838a8b0d8de4aa86193f"} Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.722985 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.758432 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-567bb65579-79jfp"] Dec 02 01:42:50 crc kubenswrapper[4884]: E1202 01:42:50.758672 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="232b9fb4-1281-4f12-ab4e-8b00aa0d4002" containerName="controller-manager" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.758691 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="232b9fb4-1281-4f12-ab4e-8b00aa0d4002" containerName="controller-manager" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.758796 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="232b9fb4-1281-4f12-ab4e-8b00aa0d4002" containerName="controller-manager" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.759193 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.775878 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-567bb65579-79jfp"] Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.826312 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-client-ca\") pod \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.826366 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-serving-cert\") pod \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.826439 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw9z4\" (UniqueName: \"kubernetes.io/projected/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-kube-api-access-rw9z4\") pod \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.826496 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-config\") pod \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.826526 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-proxy-ca-bundles\") pod \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\" (UID: \"232b9fb4-1281-4f12-ab4e-8b00aa0d4002\") " Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.826980 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c08f604-2af0-43e0-8ca1-a5ad10176291-serving-cert\") pod \"controller-manager-567bb65579-79jfp\" (UID: \"3c08f604-2af0-43e0-8ca1-a5ad10176291\") " pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.827090 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c08f604-2af0-43e0-8ca1-a5ad10176291-config\") pod \"controller-manager-567bb65579-79jfp\" (UID: \"3c08f604-2af0-43e0-8ca1-a5ad10176291\") " pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.827113 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3c08f604-2af0-43e0-8ca1-a5ad10176291-proxy-ca-bundles\") pod \"controller-manager-567bb65579-79jfp\" (UID: \"3c08f604-2af0-43e0-8ca1-a5ad10176291\") " pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.827283 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vdvh\" (UniqueName: \"kubernetes.io/projected/3c08f604-2af0-43e0-8ca1-a5ad10176291-kube-api-access-6vdvh\") pod \"controller-manager-567bb65579-79jfp\" (UID: \"3c08f604-2af0-43e0-8ca1-a5ad10176291\") " pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.827356 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3c08f604-2af0-43e0-8ca1-a5ad10176291-client-ca\") pod \"controller-manager-567bb65579-79jfp\" (UID: \"3c08f604-2af0-43e0-8ca1-a5ad10176291\") " pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.827626 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-config" (OuterVolumeSpecName: "config") pod "232b9fb4-1281-4f12-ab4e-8b00aa0d4002" (UID: "232b9fb4-1281-4f12-ab4e-8b00aa0d4002"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.827648 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "232b9fb4-1281-4f12-ab4e-8b00aa0d4002" (UID: "232b9fb4-1281-4f12-ab4e-8b00aa0d4002"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.827680 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-client-ca" (OuterVolumeSpecName: "client-ca") pod "232b9fb4-1281-4f12-ab4e-8b00aa0d4002" (UID: "232b9fb4-1281-4f12-ab4e-8b00aa0d4002"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.833388 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "232b9fb4-1281-4f12-ab4e-8b00aa0d4002" (UID: "232b9fb4-1281-4f12-ab4e-8b00aa0d4002"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.840724 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-kube-api-access-rw9z4" (OuterVolumeSpecName: "kube-api-access-rw9z4") pod "232b9fb4-1281-4f12-ab4e-8b00aa0d4002" (UID: "232b9fb4-1281-4f12-ab4e-8b00aa0d4002"). InnerVolumeSpecName "kube-api-access-rw9z4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.928469 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c08f604-2af0-43e0-8ca1-a5ad10176291-config\") pod \"controller-manager-567bb65579-79jfp\" (UID: \"3c08f604-2af0-43e0-8ca1-a5ad10176291\") " pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.928564 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3c08f604-2af0-43e0-8ca1-a5ad10176291-proxy-ca-bundles\") pod \"controller-manager-567bb65579-79jfp\" (UID: \"3c08f604-2af0-43e0-8ca1-a5ad10176291\") " pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.928648 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vdvh\" (UniqueName: \"kubernetes.io/projected/3c08f604-2af0-43e0-8ca1-a5ad10176291-kube-api-access-6vdvh\") pod \"controller-manager-567bb65579-79jfp\" (UID: \"3c08f604-2af0-43e0-8ca1-a5ad10176291\") " pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.928725 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3c08f604-2af0-43e0-8ca1-a5ad10176291-client-ca\") pod \"controller-manager-567bb65579-79jfp\" (UID: \"3c08f604-2af0-43e0-8ca1-a5ad10176291\") " pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.928830 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c08f604-2af0-43e0-8ca1-a5ad10176291-serving-cert\") pod \"controller-manager-567bb65579-79jfp\" (UID: \"3c08f604-2af0-43e0-8ca1-a5ad10176291\") " pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.928970 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.929000 4884 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.929028 4884 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-client-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.929051 4884 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.929074 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw9z4\" (UniqueName: \"kubernetes.io/projected/232b9fb4-1281-4f12-ab4e-8b00aa0d4002-kube-api-access-rw9z4\") on node \"crc\" DevicePath \"\"" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.930554 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3c08f604-2af0-43e0-8ca1-a5ad10176291-client-ca\") pod \"controller-manager-567bb65579-79jfp\" (UID: \"3c08f604-2af0-43e0-8ca1-a5ad10176291\") " pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.930845 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c08f604-2af0-43e0-8ca1-a5ad10176291-config\") pod \"controller-manager-567bb65579-79jfp\" (UID: \"3c08f604-2af0-43e0-8ca1-a5ad10176291\") " pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.931982 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3c08f604-2af0-43e0-8ca1-a5ad10176291-proxy-ca-bundles\") pod \"controller-manager-567bb65579-79jfp\" (UID: \"3c08f604-2af0-43e0-8ca1-a5ad10176291\") " pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.934652 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3c08f604-2af0-43e0-8ca1-a5ad10176291-serving-cert\") pod \"controller-manager-567bb65579-79jfp\" (UID: \"3c08f604-2af0-43e0-8ca1-a5ad10176291\") " pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:50 crc kubenswrapper[4884]: I1202 01:42:50.946923 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vdvh\" (UniqueName: \"kubernetes.io/projected/3c08f604-2af0-43e0-8ca1-a5ad10176291-kube-api-access-6vdvh\") pod \"controller-manager-567bb65579-79jfp\" (UID: \"3c08f604-2af0-43e0-8ca1-a5ad10176291\") " pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:51 crc kubenswrapper[4884]: I1202 01:42:51.078585 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:51 crc kubenswrapper[4884]: I1202 01:42:51.367587 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-567bb65579-79jfp"] Dec 02 01:42:51 crc kubenswrapper[4884]: W1202 01:42:51.373259 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c08f604_2af0_43e0_8ca1_a5ad10176291.slice/crio-486815a333c332882064c38462aff2cd43717ebda36a7384318d51c22abd5800 WatchSource:0}: Error finding container 486815a333c332882064c38462aff2cd43717ebda36a7384318d51c22abd5800: Status 404 returned error can't find the container with id 486815a333c332882064c38462aff2cd43717ebda36a7384318d51c22abd5800 Dec 02 01:42:51 crc kubenswrapper[4884]: I1202 01:42:51.532206 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" event={"ID":"232b9fb4-1281-4f12-ab4e-8b00aa0d4002","Type":"ContainerDied","Data":"7b7ea84d9c3fc0cc820bd1791d87e10511203e838e6cbaa7d549552313709261"} Dec 02 01:42:51 crc kubenswrapper[4884]: I1202 01:42:51.532272 4884 scope.go:117] "RemoveContainer" containerID="b38d6faf09251e97809f01685c0c41d439d963eb6868838a8b0d8de4aa86193f" Dec 02 01:42:51 crc kubenswrapper[4884]: I1202 01:42:51.532222 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66f89d6466-dr6vt" Dec 02 01:42:51 crc kubenswrapper[4884]: I1202 01:42:51.534002 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" event={"ID":"3c08f604-2af0-43e0-8ca1-a5ad10176291","Type":"ContainerStarted","Data":"4634068b02e8d29c4c6970e98e31177515accd797f99c69a7e63ac32db80549f"} Dec 02 01:42:51 crc kubenswrapper[4884]: I1202 01:42:51.534023 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" event={"ID":"3c08f604-2af0-43e0-8ca1-a5ad10176291","Type":"ContainerStarted","Data":"486815a333c332882064c38462aff2cd43717ebda36a7384318d51c22abd5800"} Dec 02 01:42:51 crc kubenswrapper[4884]: I1202 01:42:51.534296 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:42:51 crc kubenswrapper[4884]: I1202 01:42:51.535895 4884 patch_prober.go:28] interesting pod/controller-manager-567bb65579-79jfp container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Dec 02 01:42:51 crc kubenswrapper[4884]: I1202 01:42:51.535944 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" podUID="3c08f604-2af0-43e0-8ca1-a5ad10176291" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Dec 02 01:42:51 crc kubenswrapper[4884]: I1202 01:42:51.560365 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" podStartSLOduration=3.560330917 podStartE2EDuration="3.560330917s" podCreationTimestamp="2025-12-02 01:42:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:42:51.551264149 +0000 UTC m=+388.227101043" watchObservedRunningTime="2025-12-02 01:42:51.560330917 +0000 UTC m=+388.236167841" Dec 02 01:42:51 crc kubenswrapper[4884]: I1202 01:42:51.568992 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-66f89d6466-dr6vt"] Dec 02 01:42:51 crc kubenswrapper[4884]: I1202 01:42:51.572071 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-66f89d6466-dr6vt"] Dec 02 01:42:51 crc kubenswrapper[4884]: I1202 01:42:51.624446 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="232b9fb4-1281-4f12-ab4e-8b00aa0d4002" path="/var/lib/kubelet/pods/232b9fb4-1281-4f12-ab4e-8b00aa0d4002/volumes" Dec 02 01:42:52 crc kubenswrapper[4884]: I1202 01:42:52.548875 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-567bb65579-79jfp" Dec 02 01:43:16 crc kubenswrapper[4884]: I1202 01:43:16.972098 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:43:16 crc kubenswrapper[4884]: I1202 01:43:16.972733 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:43:16 crc kubenswrapper[4884]: I1202 01:43:16.972847 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:43:16 crc kubenswrapper[4884]: I1202 01:43:16.973824 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e037a7c9f1b0c94849181495610a0b1c6c18fa025d1d93c587fd598c55910e9f"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 01:43:16 crc kubenswrapper[4884]: I1202 01:43:16.973954 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://e037a7c9f1b0c94849181495610a0b1c6c18fa025d1d93c587fd598c55910e9f" gracePeriod=600 Dec 02 01:43:17 crc kubenswrapper[4884]: I1202 01:43:17.731620 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="e037a7c9f1b0c94849181495610a0b1c6c18fa025d1d93c587fd598c55910e9f" exitCode=0 Dec 02 01:43:17 crc kubenswrapper[4884]: I1202 01:43:17.731713 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"e037a7c9f1b0c94849181495610a0b1c6c18fa025d1d93c587fd598c55910e9f"} Dec 02 01:43:17 crc kubenswrapper[4884]: I1202 01:43:17.732339 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"7a58823165e308a2f2aec8a5502b8d1ea05368d51c2ac26ca666638cb48eebff"} Dec 02 01:43:17 crc kubenswrapper[4884]: I1202 01:43:17.732373 4884 scope.go:117] "RemoveContainer" containerID="0cee2c64c3277aef77eb6cb3745acc3ca59c04d27daa74587a075ba06a836c2f" Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.175617 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz"] Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.178029 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.181148 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.182822 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.201346 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp9sv\" (UniqueName: \"kubernetes.io/projected/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-kube-api-access-xp9sv\") pod \"collect-profiles-29410665-fx5wz\" (UID: \"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.201887 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-secret-volume\") pod \"collect-profiles-29410665-fx5wz\" (UID: \"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.202047 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-config-volume\") pod \"collect-profiles-29410665-fx5wz\" (UID: \"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.241065 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz"] Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.303175 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-config-volume\") pod \"collect-profiles-29410665-fx5wz\" (UID: \"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.303393 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp9sv\" (UniqueName: \"kubernetes.io/projected/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-kube-api-access-xp9sv\") pod \"collect-profiles-29410665-fx5wz\" (UID: \"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.303455 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-secret-volume\") pod \"collect-profiles-29410665-fx5wz\" (UID: \"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.304863 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-config-volume\") pod \"collect-profiles-29410665-fx5wz\" (UID: \"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.313860 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-secret-volume\") pod \"collect-profiles-29410665-fx5wz\" (UID: \"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.334017 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp9sv\" (UniqueName: \"kubernetes.io/projected/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-kube-api-access-xp9sv\") pod \"collect-profiles-29410665-fx5wz\" (UID: \"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.513038 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" Dec 02 01:45:00 crc kubenswrapper[4884]: I1202 01:45:00.802159 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz"] Dec 02 01:45:01 crc kubenswrapper[4884]: I1202 01:45:01.645605 4884 generic.go:334] "Generic (PLEG): container finished" podID="55b9ef1b-0207-4a5e-97e8-8c05843fb6b4" containerID="db9281b4fb77ff1a2b2d9ee00b0e9f952c72aa088e60ae0a985a21297aa1aeb0" exitCode=0 Dec 02 01:45:01 crc kubenswrapper[4884]: I1202 01:45:01.645724 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" event={"ID":"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4","Type":"ContainerDied","Data":"db9281b4fb77ff1a2b2d9ee00b0e9f952c72aa088e60ae0a985a21297aa1aeb0"} Dec 02 01:45:01 crc kubenswrapper[4884]: I1202 01:45:01.646015 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" event={"ID":"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4","Type":"ContainerStarted","Data":"93404dcbaa7370bb52284f01a50373532162c8804eaf5a8fbbe31dc9712f9bc8"} Dec 02 01:45:02 crc kubenswrapper[4884]: I1202 01:45:02.962359 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" Dec 02 01:45:03 crc kubenswrapper[4884]: I1202 01:45:03.140842 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-config-volume\") pod \"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4\" (UID: \"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4\") " Dec 02 01:45:03 crc kubenswrapper[4884]: I1202 01:45:03.141212 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xp9sv\" (UniqueName: \"kubernetes.io/projected/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-kube-api-access-xp9sv\") pod \"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4\" (UID: \"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4\") " Dec 02 01:45:03 crc kubenswrapper[4884]: I1202 01:45:03.141261 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-secret-volume\") pod \"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4\" (UID: \"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4\") " Dec 02 01:45:03 crc kubenswrapper[4884]: I1202 01:45:03.142465 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-config-volume" (OuterVolumeSpecName: "config-volume") pod "55b9ef1b-0207-4a5e-97e8-8c05843fb6b4" (UID: "55b9ef1b-0207-4a5e-97e8-8c05843fb6b4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:45:03 crc kubenswrapper[4884]: I1202 01:45:03.150027 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "55b9ef1b-0207-4a5e-97e8-8c05843fb6b4" (UID: "55b9ef1b-0207-4a5e-97e8-8c05843fb6b4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:45:03 crc kubenswrapper[4884]: I1202 01:45:03.153994 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-kube-api-access-xp9sv" (OuterVolumeSpecName: "kube-api-access-xp9sv") pod "55b9ef1b-0207-4a5e-97e8-8c05843fb6b4" (UID: "55b9ef1b-0207-4a5e-97e8-8c05843fb6b4"). InnerVolumeSpecName "kube-api-access-xp9sv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:45:03 crc kubenswrapper[4884]: I1202 01:45:03.242721 4884 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 01:45:03 crc kubenswrapper[4884]: I1202 01:45:03.242831 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xp9sv\" (UniqueName: \"kubernetes.io/projected/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-kube-api-access-xp9sv\") on node \"crc\" DevicePath \"\"" Dec 02 01:45:03 crc kubenswrapper[4884]: I1202 01:45:03.242855 4884 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 01:45:03 crc kubenswrapper[4884]: I1202 01:45:03.665507 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" event={"ID":"55b9ef1b-0207-4a5e-97e8-8c05843fb6b4","Type":"ContainerDied","Data":"93404dcbaa7370bb52284f01a50373532162c8804eaf5a8fbbe31dc9712f9bc8"} Dec 02 01:45:03 crc kubenswrapper[4884]: I1202 01:45:03.665562 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz" Dec 02 01:45:03 crc kubenswrapper[4884]: I1202 01:45:03.665589 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93404dcbaa7370bb52284f01a50373532162c8804eaf5a8fbbe31dc9712f9bc8" Dec 02 01:45:23 crc kubenswrapper[4884]: I1202 01:45:23.866295 4884 scope.go:117] "RemoveContainer" containerID="b1e237eb5e71cd598c294c3337c8468a869a10dbf6222386fc6961e62e16157e" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.665578 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dtd56"] Dec 02 01:45:38 crc kubenswrapper[4884]: E1202 01:45:38.666364 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55b9ef1b-0207-4a5e-97e8-8c05843fb6b4" containerName="collect-profiles" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.666381 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="55b9ef1b-0207-4a5e-97e8-8c05843fb6b4" containerName="collect-profiles" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.666507 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="55b9ef1b-0207-4a5e-97e8-8c05843fb6b4" containerName="collect-profiles" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.666979 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.693549 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dtd56"] Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.772689 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b4286c3-9136-425d-9465-4eb0770b8878-trusted-ca\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.772738 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szxtp\" (UniqueName: \"kubernetes.io/projected/5b4286c3-9136-425d-9465-4eb0770b8878-kube-api-access-szxtp\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.772778 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5b4286c3-9136-425d-9465-4eb0770b8878-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.772801 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5b4286c3-9136-425d-9465-4eb0770b8878-registry-certificates\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.773050 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.773165 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b4286c3-9136-425d-9465-4eb0770b8878-bound-sa-token\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.773226 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5b4286c3-9136-425d-9465-4eb0770b8878-registry-tls\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.773248 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5b4286c3-9136-425d-9465-4eb0770b8878-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.795886 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.874099 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b4286c3-9136-425d-9465-4eb0770b8878-bound-sa-token\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.874152 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5b4286c3-9136-425d-9465-4eb0770b8878-registry-tls\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.874179 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5b4286c3-9136-425d-9465-4eb0770b8878-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.874232 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b4286c3-9136-425d-9465-4eb0770b8878-trusted-ca\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.874262 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szxtp\" (UniqueName: \"kubernetes.io/projected/5b4286c3-9136-425d-9465-4eb0770b8878-kube-api-access-szxtp\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.874297 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5b4286c3-9136-425d-9465-4eb0770b8878-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.874326 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5b4286c3-9136-425d-9465-4eb0770b8878-registry-certificates\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.874948 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5b4286c3-9136-425d-9465-4eb0770b8878-ca-trust-extracted\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.875590 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5b4286c3-9136-425d-9465-4eb0770b8878-registry-certificates\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.876145 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5b4286c3-9136-425d-9465-4eb0770b8878-trusted-ca\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.881030 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5b4286c3-9136-425d-9465-4eb0770b8878-installation-pull-secrets\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.892409 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szxtp\" (UniqueName: \"kubernetes.io/projected/5b4286c3-9136-425d-9465-4eb0770b8878-kube-api-access-szxtp\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.902928 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5b4286c3-9136-425d-9465-4eb0770b8878-bound-sa-token\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.925313 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5b4286c3-9136-425d-9465-4eb0770b8878-registry-tls\") pod \"image-registry-66df7c8f76-dtd56\" (UID: \"5b4286c3-9136-425d-9465-4eb0770b8878\") " pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:38 crc kubenswrapper[4884]: I1202 01:45:38.983192 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:39 crc kubenswrapper[4884]: I1202 01:45:39.188566 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-dtd56"] Dec 02 01:45:39 crc kubenswrapper[4884]: I1202 01:45:39.937050 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" event={"ID":"5b4286c3-9136-425d-9465-4eb0770b8878","Type":"ContainerStarted","Data":"b58432395a851ae2d415b3cafe69a86a389678858c271152e31ca4c014a32129"} Dec 02 01:45:39 crc kubenswrapper[4884]: I1202 01:45:39.937585 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" event={"ID":"5b4286c3-9136-425d-9465-4eb0770b8878","Type":"ContainerStarted","Data":"c9e87055619ed46de70680214a08ba1d9481f2f77d45ed65b796cdfb12112b0d"} Dec 02 01:45:39 crc kubenswrapper[4884]: I1202 01:45:39.937812 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:39 crc kubenswrapper[4884]: I1202 01:45:39.986957 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" podStartSLOduration=1.9869277300000001 podStartE2EDuration="1.98692773s" podCreationTimestamp="2025-12-02 01:45:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:45:39.985275399 +0000 UTC m=+556.661112323" watchObservedRunningTime="2025-12-02 01:45:39.98692773 +0000 UTC m=+556.662764654" Dec 02 01:45:46 crc kubenswrapper[4884]: I1202 01:45:46.972100 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:45:46 crc kubenswrapper[4884]: I1202 01:45:46.973954 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:45:58 crc kubenswrapper[4884]: I1202 01:45:58.992321 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-dtd56" Dec 02 01:45:59 crc kubenswrapper[4884]: I1202 01:45:59.077466 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cmk67"] Dec 02 01:46:16 crc kubenswrapper[4884]: I1202 01:46:16.972365 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:46:16 crc kubenswrapper[4884]: I1202 01:46:16.973086 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:46:23 crc kubenswrapper[4884]: I1202 01:46:23.911712 4884 scope.go:117] "RemoveContainer" containerID="1b4ae6f344dd37b71a8ac7e8775d232a89bc87ae68f0b7d6e611c5e755faa0c0" Dec 02 01:46:23 crc kubenswrapper[4884]: I1202 01:46:23.940669 4884 scope.go:117] "RemoveContainer" containerID="f570cb2595d6f895190507961e109ada053c3148e6f7213cb9cc7543a3fb422a" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.135938 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" podUID="e17db01a-f964-4380-ac3f-bcad065313e9" containerName="registry" containerID="cri-o://a86ad0320703025fc535b8991bb4c3d83c21b93817ea1a882d58ca9bfdf53693" gracePeriod=30 Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.605351 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.674653 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e17db01a-f964-4380-ac3f-bcad065313e9-trusted-ca\") pod \"e17db01a-f964-4380-ac3f-bcad065313e9\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.674724 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e17db01a-f964-4380-ac3f-bcad065313e9-registry-certificates\") pod \"e17db01a-f964-4380-ac3f-bcad065313e9\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.674795 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e17db01a-f964-4380-ac3f-bcad065313e9-installation-pull-secrets\") pod \"e17db01a-f964-4380-ac3f-bcad065313e9\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.675052 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"e17db01a-f964-4380-ac3f-bcad065313e9\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.675255 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-registry-tls\") pod \"e17db01a-f964-4380-ac3f-bcad065313e9\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.675341 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e17db01a-f964-4380-ac3f-bcad065313e9-ca-trust-extracted\") pod \"e17db01a-f964-4380-ac3f-bcad065313e9\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.675401 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-bound-sa-token\") pod \"e17db01a-f964-4380-ac3f-bcad065313e9\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.675464 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jr49c\" (UniqueName: \"kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-kube-api-access-jr49c\") pod \"e17db01a-f964-4380-ac3f-bcad065313e9\" (UID: \"e17db01a-f964-4380-ac3f-bcad065313e9\") " Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.676144 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e17db01a-f964-4380-ac3f-bcad065313e9-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "e17db01a-f964-4380-ac3f-bcad065313e9" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.677255 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e17db01a-f964-4380-ac3f-bcad065313e9-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "e17db01a-f964-4380-ac3f-bcad065313e9" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.683377 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e17db01a-f964-4380-ac3f-bcad065313e9-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "e17db01a-f964-4380-ac3f-bcad065313e9" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.683583 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-kube-api-access-jr49c" (OuterVolumeSpecName: "kube-api-access-jr49c") pod "e17db01a-f964-4380-ac3f-bcad065313e9" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9"). InnerVolumeSpecName "kube-api-access-jr49c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.684553 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "e17db01a-f964-4380-ac3f-bcad065313e9" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.684940 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "e17db01a-f964-4380-ac3f-bcad065313e9" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.690839 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "e17db01a-f964-4380-ac3f-bcad065313e9" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.701698 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e17db01a-f964-4380-ac3f-bcad065313e9-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "e17db01a-f964-4380-ac3f-bcad065313e9" (UID: "e17db01a-f964-4380-ac3f-bcad065313e9"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.778423 4884 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e17db01a-f964-4380-ac3f-bcad065313e9-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.778500 4884 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.778527 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jr49c\" (UniqueName: \"kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-kube-api-access-jr49c\") on node \"crc\" DevicePath \"\"" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.778552 4884 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e17db01a-f964-4380-ac3f-bcad065313e9-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.778572 4884 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e17db01a-f964-4380-ac3f-bcad065313e9-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.778592 4884 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e17db01a-f964-4380-ac3f-bcad065313e9-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 02 01:46:24 crc kubenswrapper[4884]: I1202 01:46:24.778609 4884 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e17db01a-f964-4380-ac3f-bcad065313e9-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 02 01:46:25 crc kubenswrapper[4884]: I1202 01:46:25.260967 4884 generic.go:334] "Generic (PLEG): container finished" podID="e17db01a-f964-4380-ac3f-bcad065313e9" containerID="a86ad0320703025fc535b8991bb4c3d83c21b93817ea1a882d58ca9bfdf53693" exitCode=0 Dec 02 01:46:25 crc kubenswrapper[4884]: I1202 01:46:25.261049 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" Dec 02 01:46:25 crc kubenswrapper[4884]: I1202 01:46:25.261054 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" event={"ID":"e17db01a-f964-4380-ac3f-bcad065313e9","Type":"ContainerDied","Data":"a86ad0320703025fc535b8991bb4c3d83c21b93817ea1a882d58ca9bfdf53693"} Dec 02 01:46:25 crc kubenswrapper[4884]: I1202 01:46:25.261472 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-cmk67" event={"ID":"e17db01a-f964-4380-ac3f-bcad065313e9","Type":"ContainerDied","Data":"442de4d2157002bc2666559a3e5775b7fb0c258d18295c719746e2be8f42687d"} Dec 02 01:46:25 crc kubenswrapper[4884]: I1202 01:46:25.261508 4884 scope.go:117] "RemoveContainer" containerID="a86ad0320703025fc535b8991bb4c3d83c21b93817ea1a882d58ca9bfdf53693" Dec 02 01:46:25 crc kubenswrapper[4884]: I1202 01:46:25.285334 4884 scope.go:117] "RemoveContainer" containerID="a86ad0320703025fc535b8991bb4c3d83c21b93817ea1a882d58ca9bfdf53693" Dec 02 01:46:25 crc kubenswrapper[4884]: E1202 01:46:25.286116 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a86ad0320703025fc535b8991bb4c3d83c21b93817ea1a882d58ca9bfdf53693\": container with ID starting with a86ad0320703025fc535b8991bb4c3d83c21b93817ea1a882d58ca9bfdf53693 not found: ID does not exist" containerID="a86ad0320703025fc535b8991bb4c3d83c21b93817ea1a882d58ca9bfdf53693" Dec 02 01:46:25 crc kubenswrapper[4884]: I1202 01:46:25.286160 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a86ad0320703025fc535b8991bb4c3d83c21b93817ea1a882d58ca9bfdf53693"} err="failed to get container status \"a86ad0320703025fc535b8991bb4c3d83c21b93817ea1a882d58ca9bfdf53693\": rpc error: code = NotFound desc = could not find container \"a86ad0320703025fc535b8991bb4c3d83c21b93817ea1a882d58ca9bfdf53693\": container with ID starting with a86ad0320703025fc535b8991bb4c3d83c21b93817ea1a882d58ca9bfdf53693 not found: ID does not exist" Dec 02 01:46:25 crc kubenswrapper[4884]: I1202 01:46:25.304384 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cmk67"] Dec 02 01:46:25 crc kubenswrapper[4884]: I1202 01:46:25.314014 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-cmk67"] Dec 02 01:46:25 crc kubenswrapper[4884]: I1202 01:46:25.625868 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e17db01a-f964-4380-ac3f-bcad065313e9" path="/var/lib/kubelet/pods/e17db01a-f964-4380-ac3f-bcad065313e9/volumes" Dec 02 01:46:46 crc kubenswrapper[4884]: I1202 01:46:46.971563 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:46:46 crc kubenswrapper[4884]: I1202 01:46:46.972371 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:46:46 crc kubenswrapper[4884]: I1202 01:46:46.972474 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:46:46 crc kubenswrapper[4884]: I1202 01:46:46.973403 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7a58823165e308a2f2aec8a5502b8d1ea05368d51c2ac26ca666638cb48eebff"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 01:46:46 crc kubenswrapper[4884]: I1202 01:46:46.973517 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://7a58823165e308a2f2aec8a5502b8d1ea05368d51c2ac26ca666638cb48eebff" gracePeriod=600 Dec 02 01:46:47 crc kubenswrapper[4884]: I1202 01:46:47.425052 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="7a58823165e308a2f2aec8a5502b8d1ea05368d51c2ac26ca666638cb48eebff" exitCode=0 Dec 02 01:46:47 crc kubenswrapper[4884]: I1202 01:46:47.425138 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"7a58823165e308a2f2aec8a5502b8d1ea05368d51c2ac26ca666638cb48eebff"} Dec 02 01:46:47 crc kubenswrapper[4884]: I1202 01:46:47.425485 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"ec6281f97479113b2e5dc03c077596d5c0e07b1ee23616b661b78bdd0a2d70c8"} Dec 02 01:46:47 crc kubenswrapper[4884]: I1202 01:46:47.425521 4884 scope.go:117] "RemoveContainer" containerID="e037a7c9f1b0c94849181495610a0b1c6c18fa025d1d93c587fd598c55910e9f" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.351970 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-4kwx7"] Dec 02 01:47:48 crc kubenswrapper[4884]: E1202 01:47:48.353016 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e17db01a-f964-4380-ac3f-bcad065313e9" containerName="registry" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.353037 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="e17db01a-f964-4380-ac3f-bcad065313e9" containerName="registry" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.353249 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="e17db01a-f964-4380-ac3f-bcad065313e9" containerName="registry" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.353946 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-4kwx7" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.356955 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.357217 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.361209 4884 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-hfw2p" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.368292 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-drxm6"] Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.369323 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-drxm6" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.372148 4884 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-z48tj" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.373186 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-6tp92"] Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.373898 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-6tp92" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.378400 4884 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-4tpbj" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.388326 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-4kwx7"] Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.401444 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-drxm6"] Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.408913 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-6tp92"] Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.451058 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwf5f\" (UniqueName: \"kubernetes.io/projected/8b4e9534-399a-4836-8111-56f26c6a1b5d-kube-api-access-wwf5f\") pod \"cert-manager-cainjector-7f985d654d-4kwx7\" (UID: \"8b4e9534-399a-4836-8111-56f26c6a1b5d\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-4kwx7" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.451470 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4tjc\" (UniqueName: \"kubernetes.io/projected/7708e2d7-d33c-4ffc-a56a-094109ba52ed-kube-api-access-n4tjc\") pod \"cert-manager-webhook-5655c58dd6-6tp92\" (UID: \"7708e2d7-d33c-4ffc-a56a-094109ba52ed\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-6tp92" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.451556 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bg78\" (UniqueName: \"kubernetes.io/projected/ece97af9-2c98-424d-b314-f886b2c0cc7b-kube-api-access-2bg78\") pod \"cert-manager-5b446d88c5-drxm6\" (UID: \"ece97af9-2c98-424d-b314-f886b2c0cc7b\") " pod="cert-manager/cert-manager-5b446d88c5-drxm6" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.552893 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bg78\" (UniqueName: \"kubernetes.io/projected/ece97af9-2c98-424d-b314-f886b2c0cc7b-kube-api-access-2bg78\") pod \"cert-manager-5b446d88c5-drxm6\" (UID: \"ece97af9-2c98-424d-b314-f886b2c0cc7b\") " pod="cert-manager/cert-manager-5b446d88c5-drxm6" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.553031 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwf5f\" (UniqueName: \"kubernetes.io/projected/8b4e9534-399a-4836-8111-56f26c6a1b5d-kube-api-access-wwf5f\") pod \"cert-manager-cainjector-7f985d654d-4kwx7\" (UID: \"8b4e9534-399a-4836-8111-56f26c6a1b5d\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-4kwx7" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.553155 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4tjc\" (UniqueName: \"kubernetes.io/projected/7708e2d7-d33c-4ffc-a56a-094109ba52ed-kube-api-access-n4tjc\") pod \"cert-manager-webhook-5655c58dd6-6tp92\" (UID: \"7708e2d7-d33c-4ffc-a56a-094109ba52ed\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-6tp92" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.585154 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bg78\" (UniqueName: \"kubernetes.io/projected/ece97af9-2c98-424d-b314-f886b2c0cc7b-kube-api-access-2bg78\") pod \"cert-manager-5b446d88c5-drxm6\" (UID: \"ece97af9-2c98-424d-b314-f886b2c0cc7b\") " pod="cert-manager/cert-manager-5b446d88c5-drxm6" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.586207 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4tjc\" (UniqueName: \"kubernetes.io/projected/7708e2d7-d33c-4ffc-a56a-094109ba52ed-kube-api-access-n4tjc\") pod \"cert-manager-webhook-5655c58dd6-6tp92\" (UID: \"7708e2d7-d33c-4ffc-a56a-094109ba52ed\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-6tp92" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.587008 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwf5f\" (UniqueName: \"kubernetes.io/projected/8b4e9534-399a-4836-8111-56f26c6a1b5d-kube-api-access-wwf5f\") pod \"cert-manager-cainjector-7f985d654d-4kwx7\" (UID: \"8b4e9534-399a-4836-8111-56f26c6a1b5d\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-4kwx7" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.670481 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-4kwx7" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.694222 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-drxm6" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.702079 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-6tp92" Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.921665 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-4kwx7"] Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.937923 4884 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 01:47:48 crc kubenswrapper[4884]: I1202 01:47:48.998255 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-drxm6"] Dec 02 01:47:49 crc kubenswrapper[4884]: W1202 01:47:49.005849 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podece97af9_2c98_424d_b314_f886b2c0cc7b.slice/crio-a9deb8666c81880d1376434242f692900b526f291cabe8a79157b9371bb4a836 WatchSource:0}: Error finding container a9deb8666c81880d1376434242f692900b526f291cabe8a79157b9371bb4a836: Status 404 returned error can't find the container with id a9deb8666c81880d1376434242f692900b526f291cabe8a79157b9371bb4a836 Dec 02 01:47:49 crc kubenswrapper[4884]: I1202 01:47:49.032607 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-6tp92"] Dec 02 01:47:49 crc kubenswrapper[4884]: W1202 01:47:49.036879 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7708e2d7_d33c_4ffc_a56a_094109ba52ed.slice/crio-00688f4fbb99a9488a08ecb419bf9a2f8ce40b48b9d8ac11b9bffd3ea7b285f8 WatchSource:0}: Error finding container 00688f4fbb99a9488a08ecb419bf9a2f8ce40b48b9d8ac11b9bffd3ea7b285f8: Status 404 returned error can't find the container with id 00688f4fbb99a9488a08ecb419bf9a2f8ce40b48b9d8ac11b9bffd3ea7b285f8 Dec 02 01:47:49 crc kubenswrapper[4884]: I1202 01:47:49.848696 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-drxm6" event={"ID":"ece97af9-2c98-424d-b314-f886b2c0cc7b","Type":"ContainerStarted","Data":"a9deb8666c81880d1376434242f692900b526f291cabe8a79157b9371bb4a836"} Dec 02 01:47:49 crc kubenswrapper[4884]: I1202 01:47:49.852864 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-6tp92" event={"ID":"7708e2d7-d33c-4ffc-a56a-094109ba52ed","Type":"ContainerStarted","Data":"00688f4fbb99a9488a08ecb419bf9a2f8ce40b48b9d8ac11b9bffd3ea7b285f8"} Dec 02 01:47:49 crc kubenswrapper[4884]: I1202 01:47:49.854543 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-4kwx7" event={"ID":"8b4e9534-399a-4836-8111-56f26c6a1b5d","Type":"ContainerStarted","Data":"a381549ae228b8a1ca9b95e3ea586ce16a5ffaa1326709c0e9fa02cd877dc87e"} Dec 02 01:47:52 crc kubenswrapper[4884]: I1202 01:47:52.881596 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-6tp92" event={"ID":"7708e2d7-d33c-4ffc-a56a-094109ba52ed","Type":"ContainerStarted","Data":"375038d521520989b4e66d64dc2752c73c67f1d363b4a804f90d1b6761a24f74"} Dec 02 01:47:52 crc kubenswrapper[4884]: I1202 01:47:52.882376 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-6tp92" Dec 02 01:47:52 crc kubenswrapper[4884]: I1202 01:47:52.885023 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-4kwx7" event={"ID":"8b4e9534-399a-4836-8111-56f26c6a1b5d","Type":"ContainerStarted","Data":"cdbaa46fcfa10b33c67918fde853b025ce7dc87162830c2c44d384f7a1ae9615"} Dec 02 01:47:52 crc kubenswrapper[4884]: I1202 01:47:52.888803 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-drxm6" event={"ID":"ece97af9-2c98-424d-b314-f886b2c0cc7b","Type":"ContainerStarted","Data":"5cbf2f2f712cb2203728d0cda5e105c428ea5927aef4a3adfb06280f81ef489b"} Dec 02 01:47:52 crc kubenswrapper[4884]: I1202 01:47:52.943308 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-6tp92" podStartSLOduration=1.351678656 podStartE2EDuration="4.943281596s" podCreationTimestamp="2025-12-02 01:47:48 +0000 UTC" firstStartedPulling="2025-12-02 01:47:49.039527893 +0000 UTC m=+685.715364777" lastFinishedPulling="2025-12-02 01:47:52.631130823 +0000 UTC m=+689.306967717" observedRunningTime="2025-12-02 01:47:52.911199338 +0000 UTC m=+689.587036242" watchObservedRunningTime="2025-12-02 01:47:52.943281596 +0000 UTC m=+689.619118490" Dec 02 01:47:52 crc kubenswrapper[4884]: I1202 01:47:52.947955 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-drxm6" podStartSLOduration=1.303191215 podStartE2EDuration="4.947943948s" podCreationTimestamp="2025-12-02 01:47:48 +0000 UTC" firstStartedPulling="2025-12-02 01:47:49.008304485 +0000 UTC m=+685.684141369" lastFinishedPulling="2025-12-02 01:47:52.653057168 +0000 UTC m=+689.328894102" observedRunningTime="2025-12-02 01:47:52.935983241 +0000 UTC m=+689.611820165" watchObservedRunningTime="2025-12-02 01:47:52.947943948 +0000 UTC m=+689.623780842" Dec 02 01:47:52 crc kubenswrapper[4884]: I1202 01:47:52.979223 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-4kwx7" podStartSLOduration=1.326160725 podStartE2EDuration="4.979191046s" podCreationTimestamp="2025-12-02 01:47:48 +0000 UTC" firstStartedPulling="2025-12-02 01:47:48.937693555 +0000 UTC m=+685.613530439" lastFinishedPulling="2025-12-02 01:47:52.590723856 +0000 UTC m=+689.266560760" observedRunningTime="2025-12-02 01:47:52.958952901 +0000 UTC m=+689.634789805" watchObservedRunningTime="2025-12-02 01:47:52.979191046 +0000 UTC m=+689.655027950" Dec 02 01:47:58 crc kubenswrapper[4884]: I1202 01:47:58.705864 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-6tp92" Dec 02 01:48:03 crc kubenswrapper[4884]: I1202 01:48:03.995581 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c74x8"] Dec 02 01:48:03 crc kubenswrapper[4884]: I1202 01:48:03.996624 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovn-controller" containerID="cri-o://79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844" gracePeriod=30 Dec 02 01:48:03 crc kubenswrapper[4884]: I1202 01:48:03.996687 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="sbdb" containerID="cri-o://52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782" gracePeriod=30 Dec 02 01:48:03 crc kubenswrapper[4884]: I1202 01:48:03.996767 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b" gracePeriod=30 Dec 02 01:48:03 crc kubenswrapper[4884]: I1202 01:48:03.996639 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="nbdb" containerID="cri-o://9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055" gracePeriod=30 Dec 02 01:48:03 crc kubenswrapper[4884]: I1202 01:48:03.996824 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="northd" containerID="cri-o://e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96" gracePeriod=30 Dec 02 01:48:03 crc kubenswrapper[4884]: I1202 01:48:03.996832 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="kube-rbac-proxy-node" containerID="cri-o://f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518" gracePeriod=30 Dec 02 01:48:03 crc kubenswrapper[4884]: I1202 01:48:03.996809 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovn-acl-logging" containerID="cri-o://7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623" gracePeriod=30 Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.068671 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" containerID="cri-o://dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2" gracePeriod=30 Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.083295 4884 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.084361 4884 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.086254 4884 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.091892 4884 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.091968 4884 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="nbdb" Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.097463 4884 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.099358 4884 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.099415 4884 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="sbdb" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.444019 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/3.log" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.447276 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovn-acl-logging/0.log" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.448134 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovn-controller/0.log" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.448694 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.522480 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4lkch"] Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.522901 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="kubecfg-setup" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.522934 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="kubecfg-setup" Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.522959 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.522974 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.522990 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="northd" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523004 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="northd" Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.523022 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovn-acl-logging" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523035 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovn-acl-logging" Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.523057 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523072 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.523092 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523105 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.523125 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="kube-rbac-proxy-ovn-metrics" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523140 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="kube-rbac-proxy-ovn-metrics" Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.523153 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovn-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523166 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovn-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.523186 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523198 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.523214 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="nbdb" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523227 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="nbdb" Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.523249 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="sbdb" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523262 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="sbdb" Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.523281 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="kube-rbac-proxy-node" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523295 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="kube-rbac-proxy-node" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523496 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523523 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="kube-rbac-proxy-ovn-metrics" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523546 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523565 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="kube-rbac-proxy-node" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523586 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovn-acl-logging" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523613 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovn-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523630 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="nbdb" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523648 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523665 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523681 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="northd" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.523700 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="sbdb" Dec 02 01:48:04 crc kubenswrapper[4884]: E1202 01:48:04.524137 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.524165 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.524334 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c50c39-2888-444e-bf55-49b091559978" containerName="ovnkube-controller" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.527330 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.589947 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-systemd\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590068 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-env-overrides\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590137 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-ovnkube-config\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590201 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-etc-openvswitch\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590255 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-node-log\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590324 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d8c50c39-2888-444e-bf55-49b091559978-ovn-node-metrics-cert\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590382 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-kubelet\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590450 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-cni-bin\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590567 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-var-lib-cni-networks-ovn-kubernetes\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590627 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-ovn\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590570 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590714 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-log-socket\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590571 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590703 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590798 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-cni-netd\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590624 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590858 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-var-lib-openvswitch\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590942 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-slash\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590971 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-systemd-units\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590995 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-run-ovn-kubernetes\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590685 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591030 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-ovnkube-script-lib\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591072 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-run-netns\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590688 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591087 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591099 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bx2hb\" (UniqueName: \"kubernetes.io/projected/d8c50c39-2888-444e-bf55-49b091559978-kube-api-access-bx2hb\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591199 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-openvswitch\") pod \"d8c50c39-2888-444e-bf55-49b091559978\" (UID: \"d8c50c39-2888-444e-bf55-49b091559978\") " Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590824 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-log-socket" (OuterVolumeSpecName: "log-socket") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590848 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590920 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590985 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-slash" (OuterVolumeSpecName: "host-slash") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591033 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591078 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591138 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591364 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591707 4884 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591738 4884 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591737 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591793 4884 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591884 4884 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591915 4884 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-log-socket\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591943 4884 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591971 4884 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.591998 4884 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-slash\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.592023 4884 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.592069 4884 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.592096 4884 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.592125 4884 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.592152 4884 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.592174 4884 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.592199 4884 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.590583 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-node-log" (OuterVolumeSpecName: "node-log") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.599555 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8c50c39-2888-444e-bf55-49b091559978-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.599592 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8c50c39-2888-444e-bf55-49b091559978-kube-api-access-bx2hb" (OuterVolumeSpecName: "kube-api-access-bx2hb") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "kube-api-access-bx2hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.612804 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "d8c50c39-2888-444e-bf55-49b091559978" (UID: "d8c50c39-2888-444e-bf55-49b091559978"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.693966 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-systemd-units\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.694041 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-node-log\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.694246 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-ovn-node-metrics-cert\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.694364 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-cni-bin\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.694429 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.694475 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-env-overrides\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.694518 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-run-systemd\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.694548 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-ovnkube-config\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.694696 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4dvr\" (UniqueName: \"kubernetes.io/projected/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-kube-api-access-z4dvr\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.694803 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-cni-netd\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.694863 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-ovnkube-script-lib\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.694904 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-run-openvswitch\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.694973 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-run-ovn\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.695007 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-run-netns\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.695042 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-kubelet\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.695079 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-run-ovn-kubernetes\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.695111 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-var-lib-openvswitch\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.695165 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-slash\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.695194 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-log-socket\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.695228 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-etc-openvswitch\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.695313 4884 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/d8c50c39-2888-444e-bf55-49b091559978-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.695362 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bx2hb\" (UniqueName: \"kubernetes.io/projected/d8c50c39-2888-444e-bf55-49b091559978-kube-api-access-bx2hb\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.695388 4884 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.695412 4884 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/d8c50c39-2888-444e-bf55-49b091559978-node-log\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.695437 4884 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/d8c50c39-2888-444e-bf55-49b091559978-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.796723 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-run-ovn\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.796894 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-run-ovn\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797133 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-run-netns\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797222 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-kubelet\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797265 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-var-lib-openvswitch\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797299 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-run-ovn-kubernetes\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797331 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-kubelet\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797358 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-slash\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797430 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-run-ovn-kubernetes\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797435 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-var-lib-openvswitch\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797468 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-log-socket\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797437 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-log-socket\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797394 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-slash\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797519 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-etc-openvswitch\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797553 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-systemd-units\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797583 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-node-log\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797633 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-ovn-node-metrics-cert\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797675 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-systemd-units\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797696 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-cni-bin\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797735 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-cni-bin\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797797 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797852 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-env-overrides\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797867 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-etc-openvswitch\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797859 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-node-log\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797900 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-run-systemd\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797914 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797933 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-ovnkube-config\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797972 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-run-systemd\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.797982 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4dvr\" (UniqueName: \"kubernetes.io/projected/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-kube-api-access-z4dvr\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.798056 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-cni-netd\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.798096 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-run-netns\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.798165 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-host-cni-netd\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.798106 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-ovnkube-script-lib\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.798239 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-run-openvswitch\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.798335 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-run-openvswitch\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.799220 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-env-overrides\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.799641 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-ovnkube-config\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.799661 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-ovnkube-script-lib\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.803575 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-ovn-node-metrics-cert\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.828497 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4dvr\" (UniqueName: \"kubernetes.io/projected/f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3-kube-api-access-z4dvr\") pod \"ovnkube-node-4lkch\" (UID: \"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3\") " pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.847046 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:04 crc kubenswrapper[4884]: W1202 01:48:04.882063 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf31a8fd4_003c_4b9e_8f96_a40f7df1f0b3.slice/crio-e60cfa1315d324c4a0a5764feb2ccc37d32b38bd4a53fa27a4b37fd49be02291 WatchSource:0}: Error finding container e60cfa1315d324c4a0a5764feb2ccc37d32b38bd4a53fa27a4b37fd49be02291: Status 404 returned error can't find the container with id e60cfa1315d324c4a0a5764feb2ccc37d32b38bd4a53fa27a4b37fd49be02291 Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.980068 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" event={"ID":"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3","Type":"ContainerStarted","Data":"e60cfa1315d324c4a0a5764feb2ccc37d32b38bd4a53fa27a4b37fd49be02291"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.987312 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovnkube-controller/3.log" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.990544 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovn-acl-logging/0.log" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.991477 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-c74x8_d8c50c39-2888-444e-bf55-49b091559978/ovn-controller/0.log" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992107 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8c50c39-2888-444e-bf55-49b091559978" containerID="dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2" exitCode=0 Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992150 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8c50c39-2888-444e-bf55-49b091559978" containerID="52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782" exitCode=0 Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992168 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8c50c39-2888-444e-bf55-49b091559978" containerID="9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055" exitCode=0 Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992185 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8c50c39-2888-444e-bf55-49b091559978" containerID="e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96" exitCode=0 Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992203 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8c50c39-2888-444e-bf55-49b091559978" containerID="3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b" exitCode=0 Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992219 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8c50c39-2888-444e-bf55-49b091559978" containerID="f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518" exitCode=0 Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992225 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992232 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8c50c39-2888-444e-bf55-49b091559978" containerID="7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623" exitCode=143 Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992337 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8c50c39-2888-444e-bf55-49b091559978" containerID="79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844" exitCode=143 Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992161 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerDied","Data":"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992439 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerDied","Data":"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992462 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerDied","Data":"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992482 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerDied","Data":"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992501 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerDied","Data":"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992520 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerDied","Data":"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992524 4884 scope.go:117] "RemoveContainer" containerID="dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2" Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992538 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992555 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992567 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992636 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992650 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992661 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992671 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992783 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992807 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992824 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerDied","Data":"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992883 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992898 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992912 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992922 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992933 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992984 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.992994 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.993004 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.994676 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.994797 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.994831 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerDied","Data":"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.994937 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.994994 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995007 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995019 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995029 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995076 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995091 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995102 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995112 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995123 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995180 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-c74x8" event={"ID":"d8c50c39-2888-444e-bf55-49b091559978","Type":"ContainerDied","Data":"d0f1bec63388ff2bd42e26b0db45f584c2c0ace4e18a79e18be00e8b0493a7cc"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995201 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995215 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995263 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995278 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995290 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995384 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995399 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995410 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995421 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844"} Dec 02 01:48:04 crc kubenswrapper[4884]: I1202 01:48:04.995470 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997"} Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.004724 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6l592_791af0d6-a201-430f-b09d-02c24a4b7503/kube-multus/2.log" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.005452 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6l592_791af0d6-a201-430f-b09d-02c24a4b7503/kube-multus/1.log" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.005529 4884 generic.go:334] "Generic (PLEG): container finished" podID="791af0d6-a201-430f-b09d-02c24a4b7503" containerID="e1c1a0f1f733644cf5808f88a04f9b1c235bcc20b998175fd39ea33a4a38e6de" exitCode=2 Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.005679 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6l592" event={"ID":"791af0d6-a201-430f-b09d-02c24a4b7503","Type":"ContainerDied","Data":"e1c1a0f1f733644cf5808f88a04f9b1c235bcc20b998175fd39ea33a4a38e6de"} Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.005809 4884 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b3581bf807ed4bcf2bc452f0ad29348534c8c44bdc2d6a24614c2ea43c6d63a3"} Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.007415 4884 scope.go:117] "RemoveContainer" containerID="e1c1a0f1f733644cf5808f88a04f9b1c235bcc20b998175fd39ea33a4a38e6de" Dec 02 01:48:05 crc kubenswrapper[4884]: E1202 01:48:05.008226 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-6l592_openshift-multus(791af0d6-a201-430f-b09d-02c24a4b7503)\"" pod="openshift-multus/multus-6l592" podUID="791af0d6-a201-430f-b09d-02c24a4b7503" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.025329 4884 scope.go:117] "RemoveContainer" containerID="5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.073094 4884 scope.go:117] "RemoveContainer" containerID="52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.074678 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c74x8"] Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.082432 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-c74x8"] Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.088347 4884 scope.go:117] "RemoveContainer" containerID="9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.153388 4884 scope.go:117] "RemoveContainer" containerID="e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.178038 4884 scope.go:117] "RemoveContainer" containerID="3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.195990 4884 scope.go:117] "RemoveContainer" containerID="f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.211033 4884 scope.go:117] "RemoveContainer" containerID="7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.229512 4884 scope.go:117] "RemoveContainer" containerID="79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.248185 4884 scope.go:117] "RemoveContainer" containerID="36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.279932 4884 scope.go:117] "RemoveContainer" containerID="dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2" Dec 02 01:48:05 crc kubenswrapper[4884]: E1202 01:48:05.280511 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2\": container with ID starting with dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2 not found: ID does not exist" containerID="dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.280551 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2"} err="failed to get container status \"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2\": rpc error: code = NotFound desc = could not find container \"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2\": container with ID starting with dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.280580 4884 scope.go:117] "RemoveContainer" containerID="5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9" Dec 02 01:48:05 crc kubenswrapper[4884]: E1202 01:48:05.281053 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\": container with ID starting with 5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9 not found: ID does not exist" containerID="5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.281122 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9"} err="failed to get container status \"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\": rpc error: code = NotFound desc = could not find container \"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\": container with ID starting with 5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.281167 4884 scope.go:117] "RemoveContainer" containerID="52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782" Dec 02 01:48:05 crc kubenswrapper[4884]: E1202 01:48:05.281577 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\": container with ID starting with 52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782 not found: ID does not exist" containerID="52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.281606 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782"} err="failed to get container status \"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\": rpc error: code = NotFound desc = could not find container \"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\": container with ID starting with 52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.281625 4884 scope.go:117] "RemoveContainer" containerID="9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055" Dec 02 01:48:05 crc kubenswrapper[4884]: E1202 01:48:05.282538 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\": container with ID starting with 9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055 not found: ID does not exist" containerID="9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.282580 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055"} err="failed to get container status \"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\": rpc error: code = NotFound desc = could not find container \"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\": container with ID starting with 9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.282604 4884 scope.go:117] "RemoveContainer" containerID="e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96" Dec 02 01:48:05 crc kubenswrapper[4884]: E1202 01:48:05.283103 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\": container with ID starting with e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96 not found: ID does not exist" containerID="e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.283159 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96"} err="failed to get container status \"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\": rpc error: code = NotFound desc = could not find container \"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\": container with ID starting with e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.283196 4884 scope.go:117] "RemoveContainer" containerID="3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b" Dec 02 01:48:05 crc kubenswrapper[4884]: E1202 01:48:05.283633 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\": container with ID starting with 3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b not found: ID does not exist" containerID="3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.283666 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b"} err="failed to get container status \"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\": rpc error: code = NotFound desc = could not find container \"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\": container with ID starting with 3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.283687 4884 scope.go:117] "RemoveContainer" containerID="f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518" Dec 02 01:48:05 crc kubenswrapper[4884]: E1202 01:48:05.284066 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\": container with ID starting with f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518 not found: ID does not exist" containerID="f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.284093 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518"} err="failed to get container status \"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\": rpc error: code = NotFound desc = could not find container \"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\": container with ID starting with f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.284111 4884 scope.go:117] "RemoveContainer" containerID="7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623" Dec 02 01:48:05 crc kubenswrapper[4884]: E1202 01:48:05.284450 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\": container with ID starting with 7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623 not found: ID does not exist" containerID="7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.284476 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623"} err="failed to get container status \"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\": rpc error: code = NotFound desc = could not find container \"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\": container with ID starting with 7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.284494 4884 scope.go:117] "RemoveContainer" containerID="79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844" Dec 02 01:48:05 crc kubenswrapper[4884]: E1202 01:48:05.285191 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\": container with ID starting with 79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844 not found: ID does not exist" containerID="79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.285247 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844"} err="failed to get container status \"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\": rpc error: code = NotFound desc = could not find container \"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\": container with ID starting with 79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.285284 4884 scope.go:117] "RemoveContainer" containerID="36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997" Dec 02 01:48:05 crc kubenswrapper[4884]: E1202 01:48:05.287012 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\": container with ID starting with 36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997 not found: ID does not exist" containerID="36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.287074 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997"} err="failed to get container status \"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\": rpc error: code = NotFound desc = could not find container \"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\": container with ID starting with 36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.287118 4884 scope.go:117] "RemoveContainer" containerID="dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.288587 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2"} err="failed to get container status \"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2\": rpc error: code = NotFound desc = could not find container \"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2\": container with ID starting with dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.288632 4884 scope.go:117] "RemoveContainer" containerID="5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.290044 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9"} err="failed to get container status \"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\": rpc error: code = NotFound desc = could not find container \"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\": container with ID starting with 5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.290083 4884 scope.go:117] "RemoveContainer" containerID="52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.290482 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782"} err="failed to get container status \"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\": rpc error: code = NotFound desc = could not find container \"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\": container with ID starting with 52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.290511 4884 scope.go:117] "RemoveContainer" containerID="9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.291018 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055"} err="failed to get container status \"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\": rpc error: code = NotFound desc = could not find container \"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\": container with ID starting with 9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.291044 4884 scope.go:117] "RemoveContainer" containerID="e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.291494 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96"} err="failed to get container status \"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\": rpc error: code = NotFound desc = could not find container \"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\": container with ID starting with e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.291532 4884 scope.go:117] "RemoveContainer" containerID="3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.291969 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b"} err="failed to get container status \"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\": rpc error: code = NotFound desc = could not find container \"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\": container with ID starting with 3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.292026 4884 scope.go:117] "RemoveContainer" containerID="f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.292397 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518"} err="failed to get container status \"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\": rpc error: code = NotFound desc = could not find container \"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\": container with ID starting with f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.292433 4884 scope.go:117] "RemoveContainer" containerID="7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.292940 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623"} err="failed to get container status \"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\": rpc error: code = NotFound desc = could not find container \"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\": container with ID starting with 7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.292971 4884 scope.go:117] "RemoveContainer" containerID="79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.293550 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844"} err="failed to get container status \"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\": rpc error: code = NotFound desc = could not find container \"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\": container with ID starting with 79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.293592 4884 scope.go:117] "RemoveContainer" containerID="36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.294266 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997"} err="failed to get container status \"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\": rpc error: code = NotFound desc = could not find container \"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\": container with ID starting with 36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.294302 4884 scope.go:117] "RemoveContainer" containerID="dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.295205 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2"} err="failed to get container status \"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2\": rpc error: code = NotFound desc = could not find container \"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2\": container with ID starting with dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.295248 4884 scope.go:117] "RemoveContainer" containerID="5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.295687 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9"} err="failed to get container status \"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\": rpc error: code = NotFound desc = could not find container \"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\": container with ID starting with 5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.295718 4884 scope.go:117] "RemoveContainer" containerID="52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.296164 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782"} err="failed to get container status \"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\": rpc error: code = NotFound desc = could not find container \"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\": container with ID starting with 52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.296188 4884 scope.go:117] "RemoveContainer" containerID="9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.296542 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055"} err="failed to get container status \"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\": rpc error: code = NotFound desc = could not find container \"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\": container with ID starting with 9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.296583 4884 scope.go:117] "RemoveContainer" containerID="e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.297028 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96"} err="failed to get container status \"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\": rpc error: code = NotFound desc = could not find container \"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\": container with ID starting with e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.297064 4884 scope.go:117] "RemoveContainer" containerID="3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.297738 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b"} err="failed to get container status \"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\": rpc error: code = NotFound desc = could not find container \"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\": container with ID starting with 3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.297922 4884 scope.go:117] "RemoveContainer" containerID="f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.298356 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518"} err="failed to get container status \"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\": rpc error: code = NotFound desc = could not find container \"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\": container with ID starting with f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.298403 4884 scope.go:117] "RemoveContainer" containerID="7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.298911 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623"} err="failed to get container status \"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\": rpc error: code = NotFound desc = could not find container \"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\": container with ID starting with 7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.299055 4884 scope.go:117] "RemoveContainer" containerID="79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.299576 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844"} err="failed to get container status \"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\": rpc error: code = NotFound desc = could not find container \"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\": container with ID starting with 79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.299609 4884 scope.go:117] "RemoveContainer" containerID="36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.299967 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997"} err="failed to get container status \"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\": rpc error: code = NotFound desc = could not find container \"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\": container with ID starting with 36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.299993 4884 scope.go:117] "RemoveContainer" containerID="dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.300394 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2"} err="failed to get container status \"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2\": rpc error: code = NotFound desc = could not find container \"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2\": container with ID starting with dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.300417 4884 scope.go:117] "RemoveContainer" containerID="5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.300844 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9"} err="failed to get container status \"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\": rpc error: code = NotFound desc = could not find container \"5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9\": container with ID starting with 5ed7f34338ae18a6419052def487e90ed60cd62a7e28c221758c473f3bde55d9 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.300991 4884 scope.go:117] "RemoveContainer" containerID="52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.301506 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782"} err="failed to get container status \"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\": rpc error: code = NotFound desc = could not find container \"52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782\": container with ID starting with 52853454d00e32973b5933fb0bea7965405b3a62aff1aab6d1f6aa681e6e7782 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.301719 4884 scope.go:117] "RemoveContainer" containerID="9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.302390 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055"} err="failed to get container status \"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\": rpc error: code = NotFound desc = could not find container \"9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055\": container with ID starting with 9e096ad478d9c45a695eb8a1c81808d93ac0131c0e48714acd8c4d8734a16055 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.302437 4884 scope.go:117] "RemoveContainer" containerID="e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.302971 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96"} err="failed to get container status \"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\": rpc error: code = NotFound desc = could not find container \"e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96\": container with ID starting with e47e6b4910fb3d42a3fa29aa63b1995f097c2ed56d559fd628e73b5cf3294e96 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.303119 4884 scope.go:117] "RemoveContainer" containerID="3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.303687 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b"} err="failed to get container status \"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\": rpc error: code = NotFound desc = could not find container \"3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b\": container with ID starting with 3f58383c9dd79285ac11a304066b8dc5b9dd320db52d1f22968eff9fba39607b not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.303713 4884 scope.go:117] "RemoveContainer" containerID="f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.304304 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518"} err="failed to get container status \"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\": rpc error: code = NotFound desc = could not find container \"f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518\": container with ID starting with f40544ffdec600e6b886e96078c6102c5fbeab1bbc8f14c1c4dce6fa11da9518 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.304376 4884 scope.go:117] "RemoveContainer" containerID="7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.304995 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623"} err="failed to get container status \"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\": rpc error: code = NotFound desc = could not find container \"7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623\": container with ID starting with 7d0612db5be81d6b64dec8ccfa0cfa259cca2dc6ba53340fe83bb9348a99e623 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.305137 4884 scope.go:117] "RemoveContainer" containerID="79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.305794 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844"} err="failed to get container status \"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\": rpc error: code = NotFound desc = could not find container \"79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844\": container with ID starting with 79f920e8ce5bef75711559ee010651161a3a0b9a82514e2208bf2f5f6851f844 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.305839 4884 scope.go:117] "RemoveContainer" containerID="36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.306258 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997"} err="failed to get container status \"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\": rpc error: code = NotFound desc = could not find container \"36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997\": container with ID starting with 36daf711ecf5382032e071321e2e3d1277ffdc5dfa6b530e019a2df087d0d997 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.306283 4884 scope.go:117] "RemoveContainer" containerID="dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.306711 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2"} err="failed to get container status \"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2\": rpc error: code = NotFound desc = could not find container \"dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2\": container with ID starting with dd41d0c27643e4a89d8f722a259d36899ca2f0071b073a01b7c53c88db60dad2 not found: ID does not exist" Dec 02 01:48:05 crc kubenswrapper[4884]: I1202 01:48:05.625597 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8c50c39-2888-444e-bf55-49b091559978" path="/var/lib/kubelet/pods/d8c50c39-2888-444e-bf55-49b091559978/volumes" Dec 02 01:48:06 crc kubenswrapper[4884]: I1202 01:48:06.015175 4884 generic.go:334] "Generic (PLEG): container finished" podID="f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3" containerID="72ea06b02bc3c410f009c51cf6f63ac1e4fecef3ae7048b3f9443c3eaac0f850" exitCode=0 Dec 02 01:48:06 crc kubenswrapper[4884]: I1202 01:48:06.015272 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" event={"ID":"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3","Type":"ContainerDied","Data":"72ea06b02bc3c410f009c51cf6f63ac1e4fecef3ae7048b3f9443c3eaac0f850"} Dec 02 01:48:07 crc kubenswrapper[4884]: I1202 01:48:07.033280 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" event={"ID":"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3","Type":"ContainerStarted","Data":"418c6f0f8146da589607eba2130cf38c81909049b404cd9ce6c762bffd0176c7"} Dec 02 01:48:07 crc kubenswrapper[4884]: I1202 01:48:07.033823 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" event={"ID":"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3","Type":"ContainerStarted","Data":"b3bf460692e1988391660356ddb2594b566770714a89e5e6af6bf911a090a52d"} Dec 02 01:48:07 crc kubenswrapper[4884]: I1202 01:48:07.033840 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" event={"ID":"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3","Type":"ContainerStarted","Data":"33b37b33d6c520e9ea85429563dd3c733ca2a1985d9f185aa028d4306c09f062"} Dec 02 01:48:07 crc kubenswrapper[4884]: I1202 01:48:07.033853 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" event={"ID":"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3","Type":"ContainerStarted","Data":"9feb1791f4bacec0cec64e090f0cdde03c2806d5672a41a73ef0ebd341101e06"} Dec 02 01:48:07 crc kubenswrapper[4884]: I1202 01:48:07.033871 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" event={"ID":"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3","Type":"ContainerStarted","Data":"7dbf928807087f3092c51e859be320ac278449682d24653a2f88abb962a7f1c6"} Dec 02 01:48:07 crc kubenswrapper[4884]: I1202 01:48:07.033885 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" event={"ID":"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3","Type":"ContainerStarted","Data":"33beab98cf1143b1ca157bdcb72ee5a3a75b3e43603ed0a99021ff79532cf6d9"} Dec 02 01:48:10 crc kubenswrapper[4884]: I1202 01:48:10.057453 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" event={"ID":"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3","Type":"ContainerStarted","Data":"bdaf788f187c69c8451c548b52a3698fdf08b55f948dfe675ab81f63424460e8"} Dec 02 01:48:12 crc kubenswrapper[4884]: I1202 01:48:12.078477 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" event={"ID":"f31a8fd4-003c-4b9e-8f96-a40f7df1f0b3","Type":"ContainerStarted","Data":"46123c1cf5e800b750f8f17a14cf1f973b0f3bf59c564de1bebb11e6502a4d1c"} Dec 02 01:48:12 crc kubenswrapper[4884]: I1202 01:48:12.078927 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:12 crc kubenswrapper[4884]: I1202 01:48:12.078973 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:12 crc kubenswrapper[4884]: I1202 01:48:12.078987 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:12 crc kubenswrapper[4884]: I1202 01:48:12.122378 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" podStartSLOduration=8.122358289 podStartE2EDuration="8.122358289s" podCreationTimestamp="2025-12-02 01:48:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:48:12.110047834 +0000 UTC m=+708.785884798" watchObservedRunningTime="2025-12-02 01:48:12.122358289 +0000 UTC m=+708.798195183" Dec 02 01:48:12 crc kubenswrapper[4884]: I1202 01:48:12.125461 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:12 crc kubenswrapper[4884]: I1202 01:48:12.138989 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:16 crc kubenswrapper[4884]: I1202 01:48:16.614588 4884 scope.go:117] "RemoveContainer" containerID="e1c1a0f1f733644cf5808f88a04f9b1c235bcc20b998175fd39ea33a4a38e6de" Dec 02 01:48:16 crc kubenswrapper[4884]: E1202 01:48:16.615925 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-6l592_openshift-multus(791af0d6-a201-430f-b09d-02c24a4b7503)\"" pod="openshift-multus/multus-6l592" podUID="791af0d6-a201-430f-b09d-02c24a4b7503" Dec 02 01:48:24 crc kubenswrapper[4884]: I1202 01:48:24.068957 4884 scope.go:117] "RemoveContainer" containerID="b3581bf807ed4bcf2bc452f0ad29348534c8c44bdc2d6a24614c2ea43c6d63a3" Dec 02 01:48:24 crc kubenswrapper[4884]: I1202 01:48:24.167079 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6l592_791af0d6-a201-430f-b09d-02c24a4b7503/kube-multus/2.log" Dec 02 01:48:29 crc kubenswrapper[4884]: I1202 01:48:29.940021 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm"] Dec 02 01:48:29 crc kubenswrapper[4884]: I1202 01:48:29.941574 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:29 crc kubenswrapper[4884]: I1202 01:48:29.944321 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 02 01:48:29 crc kubenswrapper[4884]: I1202 01:48:29.957156 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm"] Dec 02 01:48:30 crc kubenswrapper[4884]: I1202 01:48:30.105887 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f4a3aaad-0061-494c-a330-c7484a1a5108-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm\" (UID: \"f4a3aaad-0061-494c-a330-c7484a1a5108\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:30 crc kubenswrapper[4884]: I1202 01:48:30.106020 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f4a3aaad-0061-494c-a330-c7484a1a5108-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm\" (UID: \"f4a3aaad-0061-494c-a330-c7484a1a5108\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:30 crc kubenswrapper[4884]: I1202 01:48:30.106129 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9hgr\" (UniqueName: \"kubernetes.io/projected/f4a3aaad-0061-494c-a330-c7484a1a5108-kube-api-access-v9hgr\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm\" (UID: \"f4a3aaad-0061-494c-a330-c7484a1a5108\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:30 crc kubenswrapper[4884]: I1202 01:48:30.207177 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9hgr\" (UniqueName: \"kubernetes.io/projected/f4a3aaad-0061-494c-a330-c7484a1a5108-kube-api-access-v9hgr\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm\" (UID: \"f4a3aaad-0061-494c-a330-c7484a1a5108\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:30 crc kubenswrapper[4884]: I1202 01:48:30.207277 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f4a3aaad-0061-494c-a330-c7484a1a5108-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm\" (UID: \"f4a3aaad-0061-494c-a330-c7484a1a5108\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:30 crc kubenswrapper[4884]: I1202 01:48:30.207357 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f4a3aaad-0061-494c-a330-c7484a1a5108-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm\" (UID: \"f4a3aaad-0061-494c-a330-c7484a1a5108\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:30 crc kubenswrapper[4884]: I1202 01:48:30.208149 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f4a3aaad-0061-494c-a330-c7484a1a5108-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm\" (UID: \"f4a3aaad-0061-494c-a330-c7484a1a5108\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:30 crc kubenswrapper[4884]: I1202 01:48:30.208249 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f4a3aaad-0061-494c-a330-c7484a1a5108-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm\" (UID: \"f4a3aaad-0061-494c-a330-c7484a1a5108\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:30 crc kubenswrapper[4884]: I1202 01:48:30.239401 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9hgr\" (UniqueName: \"kubernetes.io/projected/f4a3aaad-0061-494c-a330-c7484a1a5108-kube-api-access-v9hgr\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm\" (UID: \"f4a3aaad-0061-494c-a330-c7484a1a5108\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:30 crc kubenswrapper[4884]: I1202 01:48:30.263664 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:30 crc kubenswrapper[4884]: E1202 01:48:30.303879 4884 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_openshift-marketplace_f4a3aaad-0061-494c-a330-c7484a1a5108_0(5085bfc79ac768e236413ec0406abad4c474ed1b6de76a1b0bdfb1f320ea6fab): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 02 01:48:30 crc kubenswrapper[4884]: E1202 01:48:30.304004 4884 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_openshift-marketplace_f4a3aaad-0061-494c-a330-c7484a1a5108_0(5085bfc79ac768e236413ec0406abad4c474ed1b6de76a1b0bdfb1f320ea6fab): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:30 crc kubenswrapper[4884]: E1202 01:48:30.304053 4884 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_openshift-marketplace_f4a3aaad-0061-494c-a330-c7484a1a5108_0(5085bfc79ac768e236413ec0406abad4c474ed1b6de76a1b0bdfb1f320ea6fab): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:30 crc kubenswrapper[4884]: E1202 01:48:30.304144 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_openshift-marketplace(f4a3aaad-0061-494c-a330-c7484a1a5108)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_openshift-marketplace(f4a3aaad-0061-494c-a330-c7484a1a5108)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_openshift-marketplace_f4a3aaad-0061-494c-a330-c7484a1a5108_0(5085bfc79ac768e236413ec0406abad4c474ed1b6de76a1b0bdfb1f320ea6fab): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" podUID="f4a3aaad-0061-494c-a330-c7484a1a5108" Dec 02 01:48:30 crc kubenswrapper[4884]: I1202 01:48:30.615525 4884 scope.go:117] "RemoveContainer" containerID="e1c1a0f1f733644cf5808f88a04f9b1c235bcc20b998175fd39ea33a4a38e6de" Dec 02 01:48:31 crc kubenswrapper[4884]: I1202 01:48:31.216482 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6l592_791af0d6-a201-430f-b09d-02c24a4b7503/kube-multus/2.log" Dec 02 01:48:31 crc kubenswrapper[4884]: I1202 01:48:31.216598 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:31 crc kubenswrapper[4884]: I1202 01:48:31.216905 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6l592" event={"ID":"791af0d6-a201-430f-b09d-02c24a4b7503","Type":"ContainerStarted","Data":"a1b15317cda6d4f5c814727467a20f97cb4baa9b4edccab22a95195a0ce1c3ce"} Dec 02 01:48:31 crc kubenswrapper[4884]: I1202 01:48:31.217154 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:31 crc kubenswrapper[4884]: E1202 01:48:31.251321 4884 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_openshift-marketplace_f4a3aaad-0061-494c-a330-c7484a1a5108_0(2909e2188b0d92abbbadcdb73c956613b0fc0d62fa52b545e62c6e3bcd62b72d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 02 01:48:31 crc kubenswrapper[4884]: E1202 01:48:31.251418 4884 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_openshift-marketplace_f4a3aaad-0061-494c-a330-c7484a1a5108_0(2909e2188b0d92abbbadcdb73c956613b0fc0d62fa52b545e62c6e3bcd62b72d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:31 crc kubenswrapper[4884]: E1202 01:48:31.251457 4884 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_openshift-marketplace_f4a3aaad-0061-494c-a330-c7484a1a5108_0(2909e2188b0d92abbbadcdb73c956613b0fc0d62fa52b545e62c6e3bcd62b72d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:31 crc kubenswrapper[4884]: E1202 01:48:31.251523 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_openshift-marketplace(f4a3aaad-0061-494c-a330-c7484a1a5108)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_openshift-marketplace(f4a3aaad-0061-494c-a330-c7484a1a5108)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_openshift-marketplace_f4a3aaad-0061-494c-a330-c7484a1a5108_0(2909e2188b0d92abbbadcdb73c956613b0fc0d62fa52b545e62c6e3bcd62b72d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" podUID="f4a3aaad-0061-494c-a330-c7484a1a5108" Dec 02 01:48:34 crc kubenswrapper[4884]: I1202 01:48:34.881498 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4lkch" Dec 02 01:48:43 crc kubenswrapper[4884]: I1202 01:48:43.614099 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:43 crc kubenswrapper[4884]: I1202 01:48:43.617820 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:43 crc kubenswrapper[4884]: I1202 01:48:43.872502 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm"] Dec 02 01:48:44 crc kubenswrapper[4884]: I1202 01:48:44.529493 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" event={"ID":"f4a3aaad-0061-494c-a330-c7484a1a5108","Type":"ContainerStarted","Data":"741df9c47db355ba8c66e3070f5bb69630d86090758b4488850afc7b9717bf03"} Dec 02 01:48:45 crc kubenswrapper[4884]: I1202 01:48:45.535368 4884 generic.go:334] "Generic (PLEG): container finished" podID="f4a3aaad-0061-494c-a330-c7484a1a5108" containerID="78d6cf9d33e21f768416d72129d43651762253c0b3c70e451992c8e7c46741bc" exitCode=0 Dec 02 01:48:45 crc kubenswrapper[4884]: I1202 01:48:45.535454 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" event={"ID":"f4a3aaad-0061-494c-a330-c7484a1a5108","Type":"ContainerDied","Data":"78d6cf9d33e21f768416d72129d43651762253c0b3c70e451992c8e7c46741bc"} Dec 02 01:48:47 crc kubenswrapper[4884]: I1202 01:48:47.552099 4884 generic.go:334] "Generic (PLEG): container finished" podID="f4a3aaad-0061-494c-a330-c7484a1a5108" containerID="5c45d87523d5493e04d58ec2d96be3b8f4ffa70a04401f52d33fc73182b666c0" exitCode=0 Dec 02 01:48:47 crc kubenswrapper[4884]: I1202 01:48:47.552154 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" event={"ID":"f4a3aaad-0061-494c-a330-c7484a1a5108","Type":"ContainerDied","Data":"5c45d87523d5493e04d58ec2d96be3b8f4ffa70a04401f52d33fc73182b666c0"} Dec 02 01:48:48 crc kubenswrapper[4884]: I1202 01:48:48.563781 4884 generic.go:334] "Generic (PLEG): container finished" podID="f4a3aaad-0061-494c-a330-c7484a1a5108" containerID="d092c37a5b20a7b863acaac316b5700c135dc122332217f6e9c3b597e86b4c65" exitCode=0 Dec 02 01:48:48 crc kubenswrapper[4884]: I1202 01:48:48.563851 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" event={"ID":"f4a3aaad-0061-494c-a330-c7484a1a5108","Type":"ContainerDied","Data":"d092c37a5b20a7b863acaac316b5700c135dc122332217f6e9c3b597e86b4c65"} Dec 02 01:48:49 crc kubenswrapper[4884]: I1202 01:48:49.812575 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:49 crc kubenswrapper[4884]: I1202 01:48:49.917839 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f4a3aaad-0061-494c-a330-c7484a1a5108-util\") pod \"f4a3aaad-0061-494c-a330-c7484a1a5108\" (UID: \"f4a3aaad-0061-494c-a330-c7484a1a5108\") " Dec 02 01:48:49 crc kubenswrapper[4884]: I1202 01:48:49.918182 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9hgr\" (UniqueName: \"kubernetes.io/projected/f4a3aaad-0061-494c-a330-c7484a1a5108-kube-api-access-v9hgr\") pod \"f4a3aaad-0061-494c-a330-c7484a1a5108\" (UID: \"f4a3aaad-0061-494c-a330-c7484a1a5108\") " Dec 02 01:48:49 crc kubenswrapper[4884]: I1202 01:48:49.918213 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f4a3aaad-0061-494c-a330-c7484a1a5108-bundle\") pod \"f4a3aaad-0061-494c-a330-c7484a1a5108\" (UID: \"f4a3aaad-0061-494c-a330-c7484a1a5108\") " Dec 02 01:48:49 crc kubenswrapper[4884]: I1202 01:48:49.920506 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4a3aaad-0061-494c-a330-c7484a1a5108-bundle" (OuterVolumeSpecName: "bundle") pod "f4a3aaad-0061-494c-a330-c7484a1a5108" (UID: "f4a3aaad-0061-494c-a330-c7484a1a5108"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:48:49 crc kubenswrapper[4884]: I1202 01:48:49.924583 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4a3aaad-0061-494c-a330-c7484a1a5108-kube-api-access-v9hgr" (OuterVolumeSpecName: "kube-api-access-v9hgr") pod "f4a3aaad-0061-494c-a330-c7484a1a5108" (UID: "f4a3aaad-0061-494c-a330-c7484a1a5108"). InnerVolumeSpecName "kube-api-access-v9hgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:48:49 crc kubenswrapper[4884]: I1202 01:48:49.938370 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4a3aaad-0061-494c-a330-c7484a1a5108-util" (OuterVolumeSpecName: "util") pod "f4a3aaad-0061-494c-a330-c7484a1a5108" (UID: "f4a3aaad-0061-494c-a330-c7484a1a5108"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:48:50 crc kubenswrapper[4884]: I1202 01:48:50.020511 4884 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f4a3aaad-0061-494c-a330-c7484a1a5108-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:50 crc kubenswrapper[4884]: I1202 01:48:50.020555 4884 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f4a3aaad-0061-494c-a330-c7484a1a5108-util\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:50 crc kubenswrapper[4884]: I1202 01:48:50.020579 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9hgr\" (UniqueName: \"kubernetes.io/projected/f4a3aaad-0061-494c-a330-c7484a1a5108-kube-api-access-v9hgr\") on node \"crc\" DevicePath \"\"" Dec 02 01:48:50 crc kubenswrapper[4884]: I1202 01:48:50.582811 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" event={"ID":"f4a3aaad-0061-494c-a330-c7484a1a5108","Type":"ContainerDied","Data":"741df9c47db355ba8c66e3070f5bb69630d86090758b4488850afc7b9717bf03"} Dec 02 01:48:50 crc kubenswrapper[4884]: I1202 01:48:50.582873 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="741df9c47db355ba8c66e3070f5bb69630d86090758b4488850afc7b9717bf03" Dec 02 01:48:50 crc kubenswrapper[4884]: I1202 01:48:50.582900 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm" Dec 02 01:48:53 crc kubenswrapper[4884]: I1202 01:48:53.971866 4884 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.054773 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-fbzpt"] Dec 02 01:49:03 crc kubenswrapper[4884]: E1202 01:49:03.055982 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a3aaad-0061-494c-a330-c7484a1a5108" containerName="util" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.056002 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a3aaad-0061-494c-a330-c7484a1a5108" containerName="util" Dec 02 01:49:03 crc kubenswrapper[4884]: E1202 01:49:03.056014 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a3aaad-0061-494c-a330-c7484a1a5108" containerName="pull" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.056023 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a3aaad-0061-494c-a330-c7484a1a5108" containerName="pull" Dec 02 01:49:03 crc kubenswrapper[4884]: E1202 01:49:03.056051 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4a3aaad-0061-494c-a330-c7484a1a5108" containerName="extract" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.056060 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4a3aaad-0061-494c-a330-c7484a1a5108" containerName="extract" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.056211 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4a3aaad-0061-494c-a330-c7484a1a5108" containerName="extract" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.056892 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fbzpt" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.059266 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.059284 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-sglrh" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.059328 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.068576 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-fbzpt"] Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.167515 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-9m7lr"] Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.168704 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-9m7lr" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.170467 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.170491 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-qr2ps" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.177438 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-lclj9"] Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.178281 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-lclj9" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.182649 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-9m7lr"] Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.195080 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-lclj9"] Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.200730 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d443ca5c-ecf4-49ba-97fc-faccc445af05-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86546477d5-lclj9\" (UID: \"d443ca5c-ecf4-49ba-97fc-faccc445af05\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-lclj9" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.200782 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6175052e-784f-4b4a-8fc0-7f26d6750a37-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86546477d5-9m7lr\" (UID: \"6175052e-784f-4b4a-8fc0-7f26d6750a37\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-9m7lr" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.200809 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d443ca5c-ecf4-49ba-97fc-faccc445af05-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86546477d5-lclj9\" (UID: \"d443ca5c-ecf4-49ba-97fc-faccc445af05\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-lclj9" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.200902 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk8mm\" (UniqueName: \"kubernetes.io/projected/d1dc28c3-894b-4a33-98fa-bc56f4ba69e7-kube-api-access-fk8mm\") pod \"obo-prometheus-operator-668cf9dfbb-fbzpt\" (UID: \"d1dc28c3-894b-4a33-98fa-bc56f4ba69e7\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fbzpt" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.200923 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6175052e-784f-4b4a-8fc0-7f26d6750a37-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86546477d5-9m7lr\" (UID: \"6175052e-784f-4b4a-8fc0-7f26d6750a37\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-9m7lr" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.302133 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk8mm\" (UniqueName: \"kubernetes.io/projected/d1dc28c3-894b-4a33-98fa-bc56f4ba69e7-kube-api-access-fk8mm\") pod \"obo-prometheus-operator-668cf9dfbb-fbzpt\" (UID: \"d1dc28c3-894b-4a33-98fa-bc56f4ba69e7\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fbzpt" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.302177 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6175052e-784f-4b4a-8fc0-7f26d6750a37-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86546477d5-9m7lr\" (UID: \"6175052e-784f-4b4a-8fc0-7f26d6750a37\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-9m7lr" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.302243 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d443ca5c-ecf4-49ba-97fc-faccc445af05-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86546477d5-lclj9\" (UID: \"d443ca5c-ecf4-49ba-97fc-faccc445af05\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-lclj9" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.302272 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6175052e-784f-4b4a-8fc0-7f26d6750a37-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86546477d5-9m7lr\" (UID: \"6175052e-784f-4b4a-8fc0-7f26d6750a37\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-9m7lr" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.302291 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d443ca5c-ecf4-49ba-97fc-faccc445af05-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86546477d5-lclj9\" (UID: \"d443ca5c-ecf4-49ba-97fc-faccc445af05\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-lclj9" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.308304 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6175052e-784f-4b4a-8fc0-7f26d6750a37-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86546477d5-9m7lr\" (UID: \"6175052e-784f-4b4a-8fc0-7f26d6750a37\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-9m7lr" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.308818 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d443ca5c-ecf4-49ba-97fc-faccc445af05-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-86546477d5-lclj9\" (UID: \"d443ca5c-ecf4-49ba-97fc-faccc445af05\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-lclj9" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.312572 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6175052e-784f-4b4a-8fc0-7f26d6750a37-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86546477d5-9m7lr\" (UID: \"6175052e-784f-4b4a-8fc0-7f26d6750a37\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-9m7lr" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.313259 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d443ca5c-ecf4-49ba-97fc-faccc445af05-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-86546477d5-lclj9\" (UID: \"d443ca5c-ecf4-49ba-97fc-faccc445af05\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-lclj9" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.322248 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-t8m6m"] Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.322934 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-t8m6m" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.325219 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-q9qv4" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.325374 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.332270 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk8mm\" (UniqueName: \"kubernetes.io/projected/d1dc28c3-894b-4a33-98fa-bc56f4ba69e7-kube-api-access-fk8mm\") pod \"obo-prometheus-operator-668cf9dfbb-fbzpt\" (UID: \"d1dc28c3-894b-4a33-98fa-bc56f4ba69e7\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fbzpt" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.338024 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-t8m6m"] Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.379722 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fbzpt" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.402661 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpkhb\" (UniqueName: \"kubernetes.io/projected/5624efce-3f02-447f-907f-4041a8a86629-kube-api-access-wpkhb\") pod \"observability-operator-d8bb48f5d-t8m6m\" (UID: \"5624efce-3f02-447f-907f-4041a8a86629\") " pod="openshift-operators/observability-operator-d8bb48f5d-t8m6m" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.402754 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/5624efce-3f02-447f-907f-4041a8a86629-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-t8m6m\" (UID: \"5624efce-3f02-447f-907f-4041a8a86629\") " pod="openshift-operators/observability-operator-d8bb48f5d-t8m6m" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.479452 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-h2v4b"] Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.484141 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-9m7lr" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.491607 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-lclj9" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.497675 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-h2v4b" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.500559 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-dvgs8" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.504541 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/a5fac10c-4a51-4406-a692-9020732cc0a4-openshift-service-ca\") pod \"perses-operator-5446b9c989-h2v4b\" (UID: \"a5fac10c-4a51-4406-a692-9020732cc0a4\") " pod="openshift-operators/perses-operator-5446b9c989-h2v4b" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.504578 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpkhb\" (UniqueName: \"kubernetes.io/projected/5624efce-3f02-447f-907f-4041a8a86629-kube-api-access-wpkhb\") pod \"observability-operator-d8bb48f5d-t8m6m\" (UID: \"5624efce-3f02-447f-907f-4041a8a86629\") " pod="openshift-operators/observability-operator-d8bb48f5d-t8m6m" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.504641 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/5624efce-3f02-447f-907f-4041a8a86629-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-t8m6m\" (UID: \"5624efce-3f02-447f-907f-4041a8a86629\") " pod="openshift-operators/observability-operator-d8bb48f5d-t8m6m" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.504663 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk8kn\" (UniqueName: \"kubernetes.io/projected/a5fac10c-4a51-4406-a692-9020732cc0a4-kube-api-access-nk8kn\") pod \"perses-operator-5446b9c989-h2v4b\" (UID: \"a5fac10c-4a51-4406-a692-9020732cc0a4\") " pod="openshift-operators/perses-operator-5446b9c989-h2v4b" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.505585 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-h2v4b"] Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.529514 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/5624efce-3f02-447f-907f-4041a8a86629-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-t8m6m\" (UID: \"5624efce-3f02-447f-907f-4041a8a86629\") " pod="openshift-operators/observability-operator-d8bb48f5d-t8m6m" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.559540 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpkhb\" (UniqueName: \"kubernetes.io/projected/5624efce-3f02-447f-907f-4041a8a86629-kube-api-access-wpkhb\") pod \"observability-operator-d8bb48f5d-t8m6m\" (UID: \"5624efce-3f02-447f-907f-4041a8a86629\") " pod="openshift-operators/observability-operator-d8bb48f5d-t8m6m" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.607346 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk8kn\" (UniqueName: \"kubernetes.io/projected/a5fac10c-4a51-4406-a692-9020732cc0a4-kube-api-access-nk8kn\") pod \"perses-operator-5446b9c989-h2v4b\" (UID: \"a5fac10c-4a51-4406-a692-9020732cc0a4\") " pod="openshift-operators/perses-operator-5446b9c989-h2v4b" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.615365 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/a5fac10c-4a51-4406-a692-9020732cc0a4-openshift-service-ca\") pod \"perses-operator-5446b9c989-h2v4b\" (UID: \"a5fac10c-4a51-4406-a692-9020732cc0a4\") " pod="openshift-operators/perses-operator-5446b9c989-h2v4b" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.616412 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/a5fac10c-4a51-4406-a692-9020732cc0a4-openshift-service-ca\") pod \"perses-operator-5446b9c989-h2v4b\" (UID: \"a5fac10c-4a51-4406-a692-9020732cc0a4\") " pod="openshift-operators/perses-operator-5446b9c989-h2v4b" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.658869 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk8kn\" (UniqueName: \"kubernetes.io/projected/a5fac10c-4a51-4406-a692-9020732cc0a4-kube-api-access-nk8kn\") pod \"perses-operator-5446b9c989-h2v4b\" (UID: \"a5fac10c-4a51-4406-a692-9020732cc0a4\") " pod="openshift-operators/perses-operator-5446b9c989-h2v4b" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.729543 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-fbzpt"] Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.748538 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-t8m6m" Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.812704 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-9m7lr"] Dec 02 01:49:03 crc kubenswrapper[4884]: I1202 01:49:03.866676 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-h2v4b" Dec 02 01:49:04 crc kubenswrapper[4884]: I1202 01:49:04.050478 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-lclj9"] Dec 02 01:49:04 crc kubenswrapper[4884]: W1202 01:49:04.068508 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd443ca5c_ecf4_49ba_97fc_faccc445af05.slice/crio-f9d045549d2ed673156a8fc84716021efd8f9f8ddf612b0bd916bf13043555c5 WatchSource:0}: Error finding container f9d045549d2ed673156a8fc84716021efd8f9f8ddf612b0bd916bf13043555c5: Status 404 returned error can't find the container with id f9d045549d2ed673156a8fc84716021efd8f9f8ddf612b0bd916bf13043555c5 Dec 02 01:49:04 crc kubenswrapper[4884]: I1202 01:49:04.175985 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-t8m6m"] Dec 02 01:49:04 crc kubenswrapper[4884]: W1202 01:49:04.178692 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5624efce_3f02_447f_907f_4041a8a86629.slice/crio-2920f831aa64d597245b4286693b1ef1df9a3c7260477284270ce3a561f5e986 WatchSource:0}: Error finding container 2920f831aa64d597245b4286693b1ef1df9a3c7260477284270ce3a561f5e986: Status 404 returned error can't find the container with id 2920f831aa64d597245b4286693b1ef1df9a3c7260477284270ce3a561f5e986 Dec 02 01:49:04 crc kubenswrapper[4884]: I1202 01:49:04.442149 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-h2v4b"] Dec 02 01:49:04 crc kubenswrapper[4884]: W1202 01:49:04.451524 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5fac10c_4a51_4406_a692_9020732cc0a4.slice/crio-40434587141b4807f80402216b22d1e3ea3273b758b92d8959ede42b2a4e1363 WatchSource:0}: Error finding container 40434587141b4807f80402216b22d1e3ea3273b758b92d8959ede42b2a4e1363: Status 404 returned error can't find the container with id 40434587141b4807f80402216b22d1e3ea3273b758b92d8959ede42b2a4e1363 Dec 02 01:49:04 crc kubenswrapper[4884]: I1202 01:49:04.675423 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-h2v4b" event={"ID":"a5fac10c-4a51-4406-a692-9020732cc0a4","Type":"ContainerStarted","Data":"40434587141b4807f80402216b22d1e3ea3273b758b92d8959ede42b2a4e1363"} Dec 02 01:49:04 crc kubenswrapper[4884]: I1202 01:49:04.676990 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fbzpt" event={"ID":"d1dc28c3-894b-4a33-98fa-bc56f4ba69e7","Type":"ContainerStarted","Data":"f4b7cac5621e34fb83dbd0d5bb6109e55ad95bfda0990da0fb9853002225bda2"} Dec 02 01:49:04 crc kubenswrapper[4884]: I1202 01:49:04.678306 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-lclj9" event={"ID":"d443ca5c-ecf4-49ba-97fc-faccc445af05","Type":"ContainerStarted","Data":"f9d045549d2ed673156a8fc84716021efd8f9f8ddf612b0bd916bf13043555c5"} Dec 02 01:49:04 crc kubenswrapper[4884]: I1202 01:49:04.679837 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-9m7lr" event={"ID":"6175052e-784f-4b4a-8fc0-7f26d6750a37","Type":"ContainerStarted","Data":"5404baf65176ff1f9ff90505661aedbbf4c6360100fd60ba65e28314252bfcde"} Dec 02 01:49:04 crc kubenswrapper[4884]: I1202 01:49:04.681035 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-t8m6m" event={"ID":"5624efce-3f02-447f-907f-4041a8a86629","Type":"ContainerStarted","Data":"2920f831aa64d597245b4286693b1ef1df9a3c7260477284270ce3a561f5e986"} Dec 02 01:49:16 crc kubenswrapper[4884]: I1202 01:49:16.971656 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:49:16 crc kubenswrapper[4884]: I1202 01:49:16.972125 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:49:18 crc kubenswrapper[4884]: I1202 01:49:18.852431 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-t8m6m" event={"ID":"5624efce-3f02-447f-907f-4041a8a86629","Type":"ContainerStarted","Data":"dae7302c3e91e27f00badb1f699b62a379a6677d26b7b3260699fafb937baa80"} Dec 02 01:49:18 crc kubenswrapper[4884]: I1202 01:49:18.852910 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-t8m6m" Dec 02 01:49:18 crc kubenswrapper[4884]: I1202 01:49:18.853956 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-h2v4b" event={"ID":"a5fac10c-4a51-4406-a692-9020732cc0a4","Type":"ContainerStarted","Data":"2f32113776fc3582e52732c4812044d4b3efe4e8a66e1d3caec67ad2ba58fd1e"} Dec 02 01:49:18 crc kubenswrapper[4884]: I1202 01:49:18.854851 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-h2v4b" Dec 02 01:49:18 crc kubenswrapper[4884]: I1202 01:49:18.856798 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fbzpt" event={"ID":"d1dc28c3-894b-4a33-98fa-bc56f4ba69e7","Type":"ContainerStarted","Data":"214dca3c53cc9ff123100b4cca8db7874b5e1428734d2082c3d5afe9f64a8164"} Dec 02 01:49:18 crc kubenswrapper[4884]: I1202 01:49:18.858923 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-t8m6m" Dec 02 01:49:18 crc kubenswrapper[4884]: I1202 01:49:18.860153 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-lclj9" event={"ID":"d443ca5c-ecf4-49ba-97fc-faccc445af05","Type":"ContainerStarted","Data":"cba07faf433d9da10847d15866a748b31b6185fa6c24f1210bb359234a4a41d0"} Dec 02 01:49:18 crc kubenswrapper[4884]: I1202 01:49:18.861571 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-9m7lr" event={"ID":"6175052e-784f-4b4a-8fc0-7f26d6750a37","Type":"ContainerStarted","Data":"da1c927bd43e79f026b14831116668fafecd5ff0d7355abc6910083e165ba1f2"} Dec 02 01:49:18 crc kubenswrapper[4884]: I1202 01:49:18.905530 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-lclj9" podStartSLOduration=2.381097594 podStartE2EDuration="15.905508929s" podCreationTimestamp="2025-12-02 01:49:03 +0000 UTC" firstStartedPulling="2025-12-02 01:49:04.073100921 +0000 UTC m=+760.748937805" lastFinishedPulling="2025-12-02 01:49:17.597512256 +0000 UTC m=+774.273349140" observedRunningTime="2025-12-02 01:49:18.903473861 +0000 UTC m=+775.579310775" watchObservedRunningTime="2025-12-02 01:49:18.905508929 +0000 UTC m=+775.581345833" Dec 02 01:49:18 crc kubenswrapper[4884]: I1202 01:49:18.906947 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-t8m6m" podStartSLOduration=2.4937086 podStartE2EDuration="15.906941474s" podCreationTimestamp="2025-12-02 01:49:03 +0000 UTC" firstStartedPulling="2025-12-02 01:49:04.182380247 +0000 UTC m=+760.858217131" lastFinishedPulling="2025-12-02 01:49:17.595613121 +0000 UTC m=+774.271450005" observedRunningTime="2025-12-02 01:49:18.886453563 +0000 UTC m=+775.562290447" watchObservedRunningTime="2025-12-02 01:49:18.906941474 +0000 UTC m=+775.582778368" Dec 02 01:49:18 crc kubenswrapper[4884]: I1202 01:49:18.971449 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-86546477d5-9m7lr" podStartSLOduration=2.235682673 podStartE2EDuration="15.971426517s" podCreationTimestamp="2025-12-02 01:49:03 +0000 UTC" firstStartedPulling="2025-12-02 01:49:03.854778085 +0000 UTC m=+760.530614959" lastFinishedPulling="2025-12-02 01:49:17.590521919 +0000 UTC m=+774.266358803" observedRunningTime="2025-12-02 01:49:18.963283482 +0000 UTC m=+775.639120366" watchObservedRunningTime="2025-12-02 01:49:18.971426517 +0000 UTC m=+775.647263401" Dec 02 01:49:19 crc kubenswrapper[4884]: I1202 01:49:19.003578 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-h2v4b" podStartSLOduration=2.869303412 podStartE2EDuration="16.003562247s" podCreationTimestamp="2025-12-02 01:49:03 +0000 UTC" firstStartedPulling="2025-12-02 01:49:04.455154577 +0000 UTC m=+761.130991461" lastFinishedPulling="2025-12-02 01:49:17.589413412 +0000 UTC m=+774.265250296" observedRunningTime="2025-12-02 01:49:18.999903249 +0000 UTC m=+775.675740133" watchObservedRunningTime="2025-12-02 01:49:19.003562247 +0000 UTC m=+775.679399131" Dec 02 01:49:19 crc kubenswrapper[4884]: I1202 01:49:19.025360 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-fbzpt" podStartSLOduration=2.201436452 podStartE2EDuration="16.025345818s" podCreationTimestamp="2025-12-02 01:49:03 +0000 UTC" firstStartedPulling="2025-12-02 01:49:03.786727515 +0000 UTC m=+760.462564399" lastFinishedPulling="2025-12-02 01:49:17.610636881 +0000 UTC m=+774.286473765" observedRunningTime="2025-12-02 01:49:19.021586608 +0000 UTC m=+775.697423502" watchObservedRunningTime="2025-12-02 01:49:19.025345818 +0000 UTC m=+775.701182702" Dec 02 01:49:23 crc kubenswrapper[4884]: I1202 01:49:23.870152 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-h2v4b" Dec 02 01:49:41 crc kubenswrapper[4884]: I1202 01:49:41.406037 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872"] Dec 02 01:49:41 crc kubenswrapper[4884]: I1202 01:49:41.407420 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" Dec 02 01:49:41 crc kubenswrapper[4884]: I1202 01:49:41.410239 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 02 01:49:41 crc kubenswrapper[4884]: I1202 01:49:41.418733 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872"] Dec 02 01:49:41 crc kubenswrapper[4884]: I1202 01:49:41.563084 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b885e1cf-71b5-4b0f-8a59-d213d14db08a-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872\" (UID: \"b885e1cf-71b5-4b0f-8a59-d213d14db08a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" Dec 02 01:49:41 crc kubenswrapper[4884]: I1202 01:49:41.563166 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tvr7\" (UniqueName: \"kubernetes.io/projected/b885e1cf-71b5-4b0f-8a59-d213d14db08a-kube-api-access-9tvr7\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872\" (UID: \"b885e1cf-71b5-4b0f-8a59-d213d14db08a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" Dec 02 01:49:41 crc kubenswrapper[4884]: I1202 01:49:41.563323 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b885e1cf-71b5-4b0f-8a59-d213d14db08a-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872\" (UID: \"b885e1cf-71b5-4b0f-8a59-d213d14db08a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" Dec 02 01:49:41 crc kubenswrapper[4884]: I1202 01:49:41.664851 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b885e1cf-71b5-4b0f-8a59-d213d14db08a-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872\" (UID: \"b885e1cf-71b5-4b0f-8a59-d213d14db08a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" Dec 02 01:49:41 crc kubenswrapper[4884]: I1202 01:49:41.664988 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b885e1cf-71b5-4b0f-8a59-d213d14db08a-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872\" (UID: \"b885e1cf-71b5-4b0f-8a59-d213d14db08a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" Dec 02 01:49:41 crc kubenswrapper[4884]: I1202 01:49:41.665040 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tvr7\" (UniqueName: \"kubernetes.io/projected/b885e1cf-71b5-4b0f-8a59-d213d14db08a-kube-api-access-9tvr7\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872\" (UID: \"b885e1cf-71b5-4b0f-8a59-d213d14db08a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" Dec 02 01:49:41 crc kubenswrapper[4884]: I1202 01:49:41.665616 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b885e1cf-71b5-4b0f-8a59-d213d14db08a-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872\" (UID: \"b885e1cf-71b5-4b0f-8a59-d213d14db08a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" Dec 02 01:49:41 crc kubenswrapper[4884]: I1202 01:49:41.665730 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b885e1cf-71b5-4b0f-8a59-d213d14db08a-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872\" (UID: \"b885e1cf-71b5-4b0f-8a59-d213d14db08a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" Dec 02 01:49:41 crc kubenswrapper[4884]: I1202 01:49:41.698370 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tvr7\" (UniqueName: \"kubernetes.io/projected/b885e1cf-71b5-4b0f-8a59-d213d14db08a-kube-api-access-9tvr7\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872\" (UID: \"b885e1cf-71b5-4b0f-8a59-d213d14db08a\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" Dec 02 01:49:41 crc kubenswrapper[4884]: I1202 01:49:41.722969 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" Dec 02 01:49:42 crc kubenswrapper[4884]: I1202 01:49:42.001404 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872"] Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.005374 4884 generic.go:334] "Generic (PLEG): container finished" podID="b885e1cf-71b5-4b0f-8a59-d213d14db08a" containerID="9d51a9d5561dbfa1dfd9b79d728aaf0d9e1a5ea83d61ff95fad88db38f4214c1" exitCode=0 Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.005468 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" event={"ID":"b885e1cf-71b5-4b0f-8a59-d213d14db08a","Type":"ContainerDied","Data":"9d51a9d5561dbfa1dfd9b79d728aaf0d9e1a5ea83d61ff95fad88db38f4214c1"} Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.007113 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" event={"ID":"b885e1cf-71b5-4b0f-8a59-d213d14db08a","Type":"ContainerStarted","Data":"4e35ab821c4be861e8ea74289bf71700fe2dc55526d5942eaf71fd6b2cd4b86b"} Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.538012 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2h2q5"] Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.540218 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.556163 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2h2q5"] Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.692316 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de559ee7-c4ba-4ebc-93f4-567083524ec8-utilities\") pod \"redhat-operators-2h2q5\" (UID: \"de559ee7-c4ba-4ebc-93f4-567083524ec8\") " pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.692458 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de559ee7-c4ba-4ebc-93f4-567083524ec8-catalog-content\") pod \"redhat-operators-2h2q5\" (UID: \"de559ee7-c4ba-4ebc-93f4-567083524ec8\") " pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.692517 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnghp\" (UniqueName: \"kubernetes.io/projected/de559ee7-c4ba-4ebc-93f4-567083524ec8-kube-api-access-xnghp\") pod \"redhat-operators-2h2q5\" (UID: \"de559ee7-c4ba-4ebc-93f4-567083524ec8\") " pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.793732 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de559ee7-c4ba-4ebc-93f4-567083524ec8-utilities\") pod \"redhat-operators-2h2q5\" (UID: \"de559ee7-c4ba-4ebc-93f4-567083524ec8\") " pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.793843 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de559ee7-c4ba-4ebc-93f4-567083524ec8-catalog-content\") pod \"redhat-operators-2h2q5\" (UID: \"de559ee7-c4ba-4ebc-93f4-567083524ec8\") " pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.793883 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnghp\" (UniqueName: \"kubernetes.io/projected/de559ee7-c4ba-4ebc-93f4-567083524ec8-kube-api-access-xnghp\") pod \"redhat-operators-2h2q5\" (UID: \"de559ee7-c4ba-4ebc-93f4-567083524ec8\") " pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.794342 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de559ee7-c4ba-4ebc-93f4-567083524ec8-catalog-content\") pod \"redhat-operators-2h2q5\" (UID: \"de559ee7-c4ba-4ebc-93f4-567083524ec8\") " pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.794680 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de559ee7-c4ba-4ebc-93f4-567083524ec8-utilities\") pod \"redhat-operators-2h2q5\" (UID: \"de559ee7-c4ba-4ebc-93f4-567083524ec8\") " pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.821350 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnghp\" (UniqueName: \"kubernetes.io/projected/de559ee7-c4ba-4ebc-93f4-567083524ec8-kube-api-access-xnghp\") pod \"redhat-operators-2h2q5\" (UID: \"de559ee7-c4ba-4ebc-93f4-567083524ec8\") " pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:43 crc kubenswrapper[4884]: I1202 01:49:43.870339 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:44 crc kubenswrapper[4884]: I1202 01:49:44.169089 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2h2q5"] Dec 02 01:49:45 crc kubenswrapper[4884]: I1202 01:49:45.019246 4884 generic.go:334] "Generic (PLEG): container finished" podID="de559ee7-c4ba-4ebc-93f4-567083524ec8" containerID="86cbd784937e15f6b908bddb8fca14161498185105be087d8bc022f9c8b03799" exitCode=0 Dec 02 01:49:45 crc kubenswrapper[4884]: I1202 01:49:45.019320 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2h2q5" event={"ID":"de559ee7-c4ba-4ebc-93f4-567083524ec8","Type":"ContainerDied","Data":"86cbd784937e15f6b908bddb8fca14161498185105be087d8bc022f9c8b03799"} Dec 02 01:49:45 crc kubenswrapper[4884]: I1202 01:49:45.019354 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2h2q5" event={"ID":"de559ee7-c4ba-4ebc-93f4-567083524ec8","Type":"ContainerStarted","Data":"205eb4a1f678c60394c6f69957e8b2de6f50b2903fc2f070a523db0569622d3c"} Dec 02 01:49:45 crc kubenswrapper[4884]: I1202 01:49:45.022163 4884 generic.go:334] "Generic (PLEG): container finished" podID="b885e1cf-71b5-4b0f-8a59-d213d14db08a" containerID="5d79b5a9e43bd8758b9244ecaabd80599a2cd96806718176abde18db78fd3675" exitCode=0 Dec 02 01:49:45 crc kubenswrapper[4884]: I1202 01:49:45.022202 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" event={"ID":"b885e1cf-71b5-4b0f-8a59-d213d14db08a","Type":"ContainerDied","Data":"5d79b5a9e43bd8758b9244ecaabd80599a2cd96806718176abde18db78fd3675"} Dec 02 01:49:46 crc kubenswrapper[4884]: I1202 01:49:46.035397 4884 generic.go:334] "Generic (PLEG): container finished" podID="b885e1cf-71b5-4b0f-8a59-d213d14db08a" containerID="6bac974b06f1716e3a14a4a180d26ff60188b23dbe55f7f45a7de5d29a84d0b3" exitCode=0 Dec 02 01:49:46 crc kubenswrapper[4884]: I1202 01:49:46.035476 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" event={"ID":"b885e1cf-71b5-4b0f-8a59-d213d14db08a","Type":"ContainerDied","Data":"6bac974b06f1716e3a14a4a180d26ff60188b23dbe55f7f45a7de5d29a84d0b3"} Dec 02 01:49:46 crc kubenswrapper[4884]: I1202 01:49:46.972272 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:49:46 crc kubenswrapper[4884]: I1202 01:49:46.972387 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:49:47 crc kubenswrapper[4884]: I1202 01:49:47.045446 4884 generic.go:334] "Generic (PLEG): container finished" podID="de559ee7-c4ba-4ebc-93f4-567083524ec8" containerID="b7a84f24899b67a103b60ca7de5ddfa92ffa404ab40cd79710879047a2a4a2ac" exitCode=0 Dec 02 01:49:47 crc kubenswrapper[4884]: I1202 01:49:47.045542 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2h2q5" event={"ID":"de559ee7-c4ba-4ebc-93f4-567083524ec8","Type":"ContainerDied","Data":"b7a84f24899b67a103b60ca7de5ddfa92ffa404ab40cd79710879047a2a4a2ac"} Dec 02 01:49:47 crc kubenswrapper[4884]: I1202 01:49:47.374918 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" Dec 02 01:49:47 crc kubenswrapper[4884]: I1202 01:49:47.541093 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b885e1cf-71b5-4b0f-8a59-d213d14db08a-bundle\") pod \"b885e1cf-71b5-4b0f-8a59-d213d14db08a\" (UID: \"b885e1cf-71b5-4b0f-8a59-d213d14db08a\") " Dec 02 01:49:47 crc kubenswrapper[4884]: I1202 01:49:47.541320 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tvr7\" (UniqueName: \"kubernetes.io/projected/b885e1cf-71b5-4b0f-8a59-d213d14db08a-kube-api-access-9tvr7\") pod \"b885e1cf-71b5-4b0f-8a59-d213d14db08a\" (UID: \"b885e1cf-71b5-4b0f-8a59-d213d14db08a\") " Dec 02 01:49:47 crc kubenswrapper[4884]: I1202 01:49:47.541431 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b885e1cf-71b5-4b0f-8a59-d213d14db08a-util\") pod \"b885e1cf-71b5-4b0f-8a59-d213d14db08a\" (UID: \"b885e1cf-71b5-4b0f-8a59-d213d14db08a\") " Dec 02 01:49:47 crc kubenswrapper[4884]: I1202 01:49:47.541973 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b885e1cf-71b5-4b0f-8a59-d213d14db08a-bundle" (OuterVolumeSpecName: "bundle") pod "b885e1cf-71b5-4b0f-8a59-d213d14db08a" (UID: "b885e1cf-71b5-4b0f-8a59-d213d14db08a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:49:47 crc kubenswrapper[4884]: I1202 01:49:47.557226 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b885e1cf-71b5-4b0f-8a59-d213d14db08a-kube-api-access-9tvr7" (OuterVolumeSpecName: "kube-api-access-9tvr7") pod "b885e1cf-71b5-4b0f-8a59-d213d14db08a" (UID: "b885e1cf-71b5-4b0f-8a59-d213d14db08a"). InnerVolumeSpecName "kube-api-access-9tvr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:49:47 crc kubenswrapper[4884]: I1202 01:49:47.651036 4884 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/b885e1cf-71b5-4b0f-8a59-d213d14db08a-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:49:47 crc kubenswrapper[4884]: I1202 01:49:47.651078 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tvr7\" (UniqueName: \"kubernetes.io/projected/b885e1cf-71b5-4b0f-8a59-d213d14db08a-kube-api-access-9tvr7\") on node \"crc\" DevicePath \"\"" Dec 02 01:49:47 crc kubenswrapper[4884]: I1202 01:49:47.732699 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b885e1cf-71b5-4b0f-8a59-d213d14db08a-util" (OuterVolumeSpecName: "util") pod "b885e1cf-71b5-4b0f-8a59-d213d14db08a" (UID: "b885e1cf-71b5-4b0f-8a59-d213d14db08a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:49:47 crc kubenswrapper[4884]: I1202 01:49:47.752333 4884 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/b885e1cf-71b5-4b0f-8a59-d213d14db08a-util\") on node \"crc\" DevicePath \"\"" Dec 02 01:49:48 crc kubenswrapper[4884]: I1202 01:49:48.054268 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" event={"ID":"b885e1cf-71b5-4b0f-8a59-d213d14db08a","Type":"ContainerDied","Data":"4e35ab821c4be861e8ea74289bf71700fe2dc55526d5942eaf71fd6b2cd4b86b"} Dec 02 01:49:48 crc kubenswrapper[4884]: I1202 01:49:48.054313 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e35ab821c4be861e8ea74289bf71700fe2dc55526d5942eaf71fd6b2cd4b86b" Dec 02 01:49:48 crc kubenswrapper[4884]: I1202 01:49:48.054382 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872" Dec 02 01:49:48 crc kubenswrapper[4884]: I1202 01:49:48.064095 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2h2q5" event={"ID":"de559ee7-c4ba-4ebc-93f4-567083524ec8","Type":"ContainerStarted","Data":"34262cdc7dd5ac2db16f9a0c35d6ec43c80d74286e9aca46b28fbc7e58c7e8c7"} Dec 02 01:49:48 crc kubenswrapper[4884]: I1202 01:49:48.422081 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2h2q5" podStartSLOduration=2.666218262 podStartE2EDuration="5.422062931s" podCreationTimestamp="2025-12-02 01:49:43 +0000 UTC" firstStartedPulling="2025-12-02 01:49:45.021000095 +0000 UTC m=+801.696836979" lastFinishedPulling="2025-12-02 01:49:47.776844754 +0000 UTC m=+804.452681648" observedRunningTime="2025-12-02 01:49:48.08959706 +0000 UTC m=+804.765433974" watchObservedRunningTime="2025-12-02 01:49:48.422062931 +0000 UTC m=+805.097899815" Dec 02 01:49:51 crc kubenswrapper[4884]: I1202 01:49:51.557066 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-wm7q4"] Dec 02 01:49:51 crc kubenswrapper[4884]: E1202 01:49:51.557284 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b885e1cf-71b5-4b0f-8a59-d213d14db08a" containerName="pull" Dec 02 01:49:51 crc kubenswrapper[4884]: I1202 01:49:51.557298 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="b885e1cf-71b5-4b0f-8a59-d213d14db08a" containerName="pull" Dec 02 01:49:51 crc kubenswrapper[4884]: E1202 01:49:51.557309 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b885e1cf-71b5-4b0f-8a59-d213d14db08a" containerName="extract" Dec 02 01:49:51 crc kubenswrapper[4884]: I1202 01:49:51.557314 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="b885e1cf-71b5-4b0f-8a59-d213d14db08a" containerName="extract" Dec 02 01:49:51 crc kubenswrapper[4884]: E1202 01:49:51.557325 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b885e1cf-71b5-4b0f-8a59-d213d14db08a" containerName="util" Dec 02 01:49:51 crc kubenswrapper[4884]: I1202 01:49:51.557331 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="b885e1cf-71b5-4b0f-8a59-d213d14db08a" containerName="util" Dec 02 01:49:51 crc kubenswrapper[4884]: I1202 01:49:51.557432 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="b885e1cf-71b5-4b0f-8a59-d213d14db08a" containerName="extract" Dec 02 01:49:51 crc kubenswrapper[4884]: I1202 01:49:51.557843 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wm7q4" Dec 02 01:49:51 crc kubenswrapper[4884]: I1202 01:49:51.560487 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 02 01:49:51 crc kubenswrapper[4884]: I1202 01:49:51.560528 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-8sj8z" Dec 02 01:49:51 crc kubenswrapper[4884]: I1202 01:49:51.560707 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 02 01:49:51 crc kubenswrapper[4884]: I1202 01:49:51.573519 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-wm7q4"] Dec 02 01:49:51 crc kubenswrapper[4884]: I1202 01:49:51.707681 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nj8f\" (UniqueName: \"kubernetes.io/projected/7f575e02-628b-4380-a9d3-126a6ababd10-kube-api-access-5nj8f\") pod \"nmstate-operator-5b5b58f5c8-wm7q4\" (UID: \"7f575e02-628b-4380-a9d3-126a6ababd10\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wm7q4" Dec 02 01:49:51 crc kubenswrapper[4884]: I1202 01:49:51.809838 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nj8f\" (UniqueName: \"kubernetes.io/projected/7f575e02-628b-4380-a9d3-126a6ababd10-kube-api-access-5nj8f\") pod \"nmstate-operator-5b5b58f5c8-wm7q4\" (UID: \"7f575e02-628b-4380-a9d3-126a6ababd10\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wm7q4" Dec 02 01:49:51 crc kubenswrapper[4884]: I1202 01:49:51.836052 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nj8f\" (UniqueName: \"kubernetes.io/projected/7f575e02-628b-4380-a9d3-126a6ababd10-kube-api-access-5nj8f\") pod \"nmstate-operator-5b5b58f5c8-wm7q4\" (UID: \"7f575e02-628b-4380-a9d3-126a6ababd10\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wm7q4" Dec 02 01:49:51 crc kubenswrapper[4884]: I1202 01:49:51.872691 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wm7q4" Dec 02 01:49:52 crc kubenswrapper[4884]: I1202 01:49:52.120675 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-wm7q4"] Dec 02 01:49:52 crc kubenswrapper[4884]: W1202 01:49:52.127273 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f575e02_628b_4380_a9d3_126a6ababd10.slice/crio-e065481146b57cf95545cad727880eba245412cdf167cafc7413bd56c7e384d3 WatchSource:0}: Error finding container e065481146b57cf95545cad727880eba245412cdf167cafc7413bd56c7e384d3: Status 404 returned error can't find the container with id e065481146b57cf95545cad727880eba245412cdf167cafc7413bd56c7e384d3 Dec 02 01:49:53 crc kubenswrapper[4884]: I1202 01:49:53.097783 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wm7q4" event={"ID":"7f575e02-628b-4380-a9d3-126a6ababd10","Type":"ContainerStarted","Data":"e065481146b57cf95545cad727880eba245412cdf167cafc7413bd56c7e384d3"} Dec 02 01:49:53 crc kubenswrapper[4884]: I1202 01:49:53.871344 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:53 crc kubenswrapper[4884]: I1202 01:49:53.871710 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:53 crc kubenswrapper[4884]: I1202 01:49:53.941060 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:54 crc kubenswrapper[4884]: I1202 01:49:54.174379 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:56 crc kubenswrapper[4884]: I1202 01:49:56.116936 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wm7q4" event={"ID":"7f575e02-628b-4380-a9d3-126a6ababd10","Type":"ContainerStarted","Data":"9a986528c8f272aeac2bd752733d761039fb3ff1833c0d1c62fd5fbc66d0d709"} Dec 02 01:49:56 crc kubenswrapper[4884]: I1202 01:49:56.149105 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-wm7q4" podStartSLOduration=1.698812995 podStartE2EDuration="5.14908666s" podCreationTimestamp="2025-12-02 01:49:51 +0000 UTC" firstStartedPulling="2025-12-02 01:49:52.131824246 +0000 UTC m=+808.807661130" lastFinishedPulling="2025-12-02 01:49:55.582097911 +0000 UTC m=+812.257934795" observedRunningTime="2025-12-02 01:49:56.142071921 +0000 UTC m=+812.817908835" watchObservedRunningTime="2025-12-02 01:49:56.14908666 +0000 UTC m=+812.824923554" Dec 02 01:49:56 crc kubenswrapper[4884]: I1202 01:49:56.523646 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2h2q5"] Dec 02 01:49:56 crc kubenswrapper[4884]: I1202 01:49:56.524172 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2h2q5" podUID="de559ee7-c4ba-4ebc-93f4-567083524ec8" containerName="registry-server" containerID="cri-o://34262cdc7dd5ac2db16f9a0c35d6ec43c80d74286e9aca46b28fbc7e58c7e8c7" gracePeriod=2 Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.084571 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-wmss4"] Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.086140 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wmss4" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.090562 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-srjwr" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.091784 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq"] Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.092833 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.094696 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.095080 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-wmss4"] Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.102892 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f85jf\" (UniqueName: \"kubernetes.io/projected/4a1eae32-edb9-42c6-9a22-62012c26bf23-kube-api-access-f85jf\") pod \"nmstate-metrics-7f946cbc9-wmss4\" (UID: \"4a1eae32-edb9-42c6-9a22-62012c26bf23\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wmss4" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.102945 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzr84\" (UniqueName: \"kubernetes.io/projected/c783fcfc-44e0-4a67-bc02-f4d0f45f10a9-kube-api-access-bzr84\") pod \"nmstate-webhook-5f6d4c5ccb-2mzbq\" (UID: \"c783fcfc-44e0-4a67-bc02-f4d0f45f10a9\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.103002 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/c783fcfc-44e0-4a67-bc02-f4d0f45f10a9-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-2mzbq\" (UID: \"c783fcfc-44e0-4a67-bc02-f4d0f45f10a9\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.111346 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq"] Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.150273 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-d5d28"] Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.151583 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.204982 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/c783fcfc-44e0-4a67-bc02-f4d0f45f10a9-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-2mzbq\" (UID: \"c783fcfc-44e0-4a67-bc02-f4d0f45f10a9\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.205317 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7jml\" (UniqueName: \"kubernetes.io/projected/51729f65-2f0f-4a8a-bf7d-1b91f706aa82-kube-api-access-s7jml\") pod \"nmstate-handler-d5d28\" (UID: \"51729f65-2f0f-4a8a-bf7d-1b91f706aa82\") " pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:49:57 crc kubenswrapper[4884]: E1202 01:49:57.205174 4884 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 02 01:49:57 crc kubenswrapper[4884]: E1202 01:49:57.205417 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c783fcfc-44e0-4a67-bc02-f4d0f45f10a9-tls-key-pair podName:c783fcfc-44e0-4a67-bc02-f4d0f45f10a9 nodeName:}" failed. No retries permitted until 2025-12-02 01:49:57.705394805 +0000 UTC m=+814.381231679 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/c783fcfc-44e0-4a67-bc02-f4d0f45f10a9-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-2mzbq" (UID: "c783fcfc-44e0-4a67-bc02-f4d0f45f10a9") : secret "openshift-nmstate-webhook" not found Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.205697 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/51729f65-2f0f-4a8a-bf7d-1b91f706aa82-ovs-socket\") pod \"nmstate-handler-d5d28\" (UID: \"51729f65-2f0f-4a8a-bf7d-1b91f706aa82\") " pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.205733 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f85jf\" (UniqueName: \"kubernetes.io/projected/4a1eae32-edb9-42c6-9a22-62012c26bf23-kube-api-access-f85jf\") pod \"nmstate-metrics-7f946cbc9-wmss4\" (UID: \"4a1eae32-edb9-42c6-9a22-62012c26bf23\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wmss4" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.205797 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzr84\" (UniqueName: \"kubernetes.io/projected/c783fcfc-44e0-4a67-bc02-f4d0f45f10a9-kube-api-access-bzr84\") pod \"nmstate-webhook-5f6d4c5ccb-2mzbq\" (UID: \"c783fcfc-44e0-4a67-bc02-f4d0f45f10a9\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.205876 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/51729f65-2f0f-4a8a-bf7d-1b91f706aa82-dbus-socket\") pod \"nmstate-handler-d5d28\" (UID: \"51729f65-2f0f-4a8a-bf7d-1b91f706aa82\") " pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.206088 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/51729f65-2f0f-4a8a-bf7d-1b91f706aa82-nmstate-lock\") pod \"nmstate-handler-d5d28\" (UID: \"51729f65-2f0f-4a8a-bf7d-1b91f706aa82\") " pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.226311 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzr84\" (UniqueName: \"kubernetes.io/projected/c783fcfc-44e0-4a67-bc02-f4d0f45f10a9-kube-api-access-bzr84\") pod \"nmstate-webhook-5f6d4c5ccb-2mzbq\" (UID: \"c783fcfc-44e0-4a67-bc02-f4d0f45f10a9\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.241346 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5"] Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.241862 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f85jf\" (UniqueName: \"kubernetes.io/projected/4a1eae32-edb9-42c6-9a22-62012c26bf23-kube-api-access-f85jf\") pod \"nmstate-metrics-7f946cbc9-wmss4\" (UID: \"4a1eae32-edb9-42c6-9a22-62012c26bf23\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wmss4" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.242243 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.247186 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.247373 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.254090 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5"] Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.254235 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-qcswm" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.306686 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c8a2aedd-caa8-4a87-9429-9443ef9c0fe4-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-6skt5\" (UID: \"c8a2aedd-caa8-4a87-9429-9443ef9c0fe4\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.306782 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/51729f65-2f0f-4a8a-bf7d-1b91f706aa82-dbus-socket\") pod \"nmstate-handler-d5d28\" (UID: \"51729f65-2f0f-4a8a-bf7d-1b91f706aa82\") " pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.306862 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/51729f65-2f0f-4a8a-bf7d-1b91f706aa82-nmstate-lock\") pod \"nmstate-handler-d5d28\" (UID: \"51729f65-2f0f-4a8a-bf7d-1b91f706aa82\") " pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.306927 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7jml\" (UniqueName: \"kubernetes.io/projected/51729f65-2f0f-4a8a-bf7d-1b91f706aa82-kube-api-access-s7jml\") pod \"nmstate-handler-d5d28\" (UID: \"51729f65-2f0f-4a8a-bf7d-1b91f706aa82\") " pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.306985 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/51729f65-2f0f-4a8a-bf7d-1b91f706aa82-ovs-socket\") pod \"nmstate-handler-d5d28\" (UID: \"51729f65-2f0f-4a8a-bf7d-1b91f706aa82\") " pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.307008 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkf8r\" (UniqueName: \"kubernetes.io/projected/c8a2aedd-caa8-4a87-9429-9443ef9c0fe4-kube-api-access-lkf8r\") pod \"nmstate-console-plugin-7fbb5f6569-6skt5\" (UID: \"c8a2aedd-caa8-4a87-9429-9443ef9c0fe4\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.307034 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c8a2aedd-caa8-4a87-9429-9443ef9c0fe4-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-6skt5\" (UID: \"c8a2aedd-caa8-4a87-9429-9443ef9c0fe4\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.307466 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/51729f65-2f0f-4a8a-bf7d-1b91f706aa82-dbus-socket\") pod \"nmstate-handler-d5d28\" (UID: \"51729f65-2f0f-4a8a-bf7d-1b91f706aa82\") " pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.307587 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/51729f65-2f0f-4a8a-bf7d-1b91f706aa82-ovs-socket\") pod \"nmstate-handler-d5d28\" (UID: \"51729f65-2f0f-4a8a-bf7d-1b91f706aa82\") " pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.307700 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/51729f65-2f0f-4a8a-bf7d-1b91f706aa82-nmstate-lock\") pod \"nmstate-handler-d5d28\" (UID: \"51729f65-2f0f-4a8a-bf7d-1b91f706aa82\") " pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.323954 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7jml\" (UniqueName: \"kubernetes.io/projected/51729f65-2f0f-4a8a-bf7d-1b91f706aa82-kube-api-access-s7jml\") pod \"nmstate-handler-d5d28\" (UID: \"51729f65-2f0f-4a8a-bf7d-1b91f706aa82\") " pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.407986 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkf8r\" (UniqueName: \"kubernetes.io/projected/c8a2aedd-caa8-4a87-9429-9443ef9c0fe4-kube-api-access-lkf8r\") pod \"nmstate-console-plugin-7fbb5f6569-6skt5\" (UID: \"c8a2aedd-caa8-4a87-9429-9443ef9c0fe4\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.408041 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c8a2aedd-caa8-4a87-9429-9443ef9c0fe4-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-6skt5\" (UID: \"c8a2aedd-caa8-4a87-9429-9443ef9c0fe4\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.408075 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c8a2aedd-caa8-4a87-9429-9443ef9c0fe4-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-6skt5\" (UID: \"c8a2aedd-caa8-4a87-9429-9443ef9c0fe4\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.409166 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/c8a2aedd-caa8-4a87-9429-9443ef9c0fe4-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-6skt5\" (UID: \"c8a2aedd-caa8-4a87-9429-9443ef9c0fe4\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.411610 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/c8a2aedd-caa8-4a87-9429-9443ef9c0fe4-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-6skt5\" (UID: \"c8a2aedd-caa8-4a87-9429-9443ef9c0fe4\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.414990 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wmss4" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.434615 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkf8r\" (UniqueName: \"kubernetes.io/projected/c8a2aedd-caa8-4a87-9429-9443ef9c0fe4-kube-api-access-lkf8r\") pod \"nmstate-console-plugin-7fbb5f6569-6skt5\" (UID: \"c8a2aedd-caa8-4a87-9429-9443ef9c0fe4\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.471390 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:49:57 crc kubenswrapper[4884]: W1202 01:49:57.498787 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51729f65_2f0f_4a8a_bf7d_1b91f706aa82.slice/crio-5c59f2b3ea949994b7161285c74f4e59970e49c7e4785eb131999d134b7abedc WatchSource:0}: Error finding container 5c59f2b3ea949994b7161285c74f4e59970e49c7e4785eb131999d134b7abedc: Status 404 returned error can't find the container with id 5c59f2b3ea949994b7161285c74f4e59970e49c7e4785eb131999d134b7abedc Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.541967 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-746c6b4b7b-nzgxl"] Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.546313 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.550381 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-746c6b4b7b-nzgxl"] Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.577170 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.712101 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw58c\" (UniqueName: \"kubernetes.io/projected/291657a1-ad1d-4b85-8717-70a240a7f8e0-kube-api-access-vw58c\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.712296 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/291657a1-ad1d-4b85-8717-70a240a7f8e0-console-serving-cert\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.712317 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/291657a1-ad1d-4b85-8717-70a240a7f8e0-service-ca\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.712343 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/291657a1-ad1d-4b85-8717-70a240a7f8e0-trusted-ca-bundle\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.712415 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/c783fcfc-44e0-4a67-bc02-f4d0f45f10a9-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-2mzbq\" (UID: \"c783fcfc-44e0-4a67-bc02-f4d0f45f10a9\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.712437 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/291657a1-ad1d-4b85-8717-70a240a7f8e0-console-oauth-config\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.712529 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/291657a1-ad1d-4b85-8717-70a240a7f8e0-oauth-serving-cert\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.712548 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/291657a1-ad1d-4b85-8717-70a240a7f8e0-console-config\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.718690 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/c783fcfc-44e0-4a67-bc02-f4d0f45f10a9-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-2mzbq\" (UID: \"c783fcfc-44e0-4a67-bc02-f4d0f45f10a9\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.739224 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.765807 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5"] Dec 02 01:49:57 crc kubenswrapper[4884]: W1202 01:49:57.768343 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8a2aedd_caa8_4a87_9429_9443ef9c0fe4.slice/crio-07fab9c4b2c274aaf7c702aff29259ef035d93b9de222e087e1e4cead5661b1d WatchSource:0}: Error finding container 07fab9c4b2c274aaf7c702aff29259ef035d93b9de222e087e1e4cead5661b1d: Status 404 returned error can't find the container with id 07fab9c4b2c274aaf7c702aff29259ef035d93b9de222e087e1e4cead5661b1d Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.816014 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw58c\" (UniqueName: \"kubernetes.io/projected/291657a1-ad1d-4b85-8717-70a240a7f8e0-kube-api-access-vw58c\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.816060 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/291657a1-ad1d-4b85-8717-70a240a7f8e0-console-serving-cert\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.816088 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/291657a1-ad1d-4b85-8717-70a240a7f8e0-service-ca\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.816121 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/291657a1-ad1d-4b85-8717-70a240a7f8e0-trusted-ca-bundle\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.816168 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/291657a1-ad1d-4b85-8717-70a240a7f8e0-console-oauth-config\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.816191 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/291657a1-ad1d-4b85-8717-70a240a7f8e0-oauth-serving-cert\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.816211 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/291657a1-ad1d-4b85-8717-70a240a7f8e0-console-config\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.817522 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/291657a1-ad1d-4b85-8717-70a240a7f8e0-console-config\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.818661 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/291657a1-ad1d-4b85-8717-70a240a7f8e0-trusted-ca-bundle\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.818668 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/291657a1-ad1d-4b85-8717-70a240a7f8e0-service-ca\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.819014 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/291657a1-ad1d-4b85-8717-70a240a7f8e0-oauth-serving-cert\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.824376 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/291657a1-ad1d-4b85-8717-70a240a7f8e0-console-serving-cert\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.826192 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/291657a1-ad1d-4b85-8717-70a240a7f8e0-console-oauth-config\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.836325 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw58c\" (UniqueName: \"kubernetes.io/projected/291657a1-ad1d-4b85-8717-70a240a7f8e0-kube-api-access-vw58c\") pod \"console-746c6b4b7b-nzgxl\" (UID: \"291657a1-ad1d-4b85-8717-70a240a7f8e0\") " pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.869931 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-wmss4"] Dec 02 01:49:57 crc kubenswrapper[4884]: W1202 01:49:57.880838 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a1eae32_edb9_42c6_9a22_62012c26bf23.slice/crio-f48d2457d1783526597f867ac6163e6ba293e324f44ec2e1787e4d00f29a9e7f WatchSource:0}: Error finding container f48d2457d1783526597f867ac6163e6ba293e324f44ec2e1787e4d00f29a9e7f: Status 404 returned error can't find the container with id f48d2457d1783526597f867ac6163e6ba293e324f44ec2e1787e4d00f29a9e7f Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.881282 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:49:57 crc kubenswrapper[4884]: I1202 01:49:57.964463 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq"] Dec 02 01:49:57 crc kubenswrapper[4884]: W1202 01:49:57.964475 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc783fcfc_44e0_4a67_bc02_f4d0f45f10a9.slice/crio-1b67b8bb9e8e8a3b7c193bbc8afeff960f989012e5749d549fc0b9b5955010fd WatchSource:0}: Error finding container 1b67b8bb9e8e8a3b7c193bbc8afeff960f989012e5749d549fc0b9b5955010fd: Status 404 returned error can't find the container with id 1b67b8bb9e8e8a3b7c193bbc8afeff960f989012e5749d549fc0b9b5955010fd Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.114256 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-746c6b4b7b-nzgxl"] Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.157713 4884 generic.go:334] "Generic (PLEG): container finished" podID="de559ee7-c4ba-4ebc-93f4-567083524ec8" containerID="34262cdc7dd5ac2db16f9a0c35d6ec43c80d74286e9aca46b28fbc7e58c7e8c7" exitCode=0 Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.157786 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2h2q5" event={"ID":"de559ee7-c4ba-4ebc-93f4-567083524ec8","Type":"ContainerDied","Data":"34262cdc7dd5ac2db16f9a0c35d6ec43c80d74286e9aca46b28fbc7e58c7e8c7"} Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.158773 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5" event={"ID":"c8a2aedd-caa8-4a87-9429-9443ef9c0fe4","Type":"ContainerStarted","Data":"07fab9c4b2c274aaf7c702aff29259ef035d93b9de222e087e1e4cead5661b1d"} Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.163017 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-746c6b4b7b-nzgxl" event={"ID":"291657a1-ad1d-4b85-8717-70a240a7f8e0","Type":"ContainerStarted","Data":"c128ed5055f44fcb87bb42d1c3b0ac9dc58335cdb8c07c00e6541689f8596683"} Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.164110 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-d5d28" event={"ID":"51729f65-2f0f-4a8a-bf7d-1b91f706aa82","Type":"ContainerStarted","Data":"5c59f2b3ea949994b7161285c74f4e59970e49c7e4785eb131999d134b7abedc"} Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.169022 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq" event={"ID":"c783fcfc-44e0-4a67-bc02-f4d0f45f10a9","Type":"ContainerStarted","Data":"1b67b8bb9e8e8a3b7c193bbc8afeff960f989012e5749d549fc0b9b5955010fd"} Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.169867 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wmss4" event={"ID":"4a1eae32-edb9-42c6-9a22-62012c26bf23","Type":"ContainerStarted","Data":"f48d2457d1783526597f867ac6163e6ba293e324f44ec2e1787e4d00f29a9e7f"} Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.436868 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.537599 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnghp\" (UniqueName: \"kubernetes.io/projected/de559ee7-c4ba-4ebc-93f4-567083524ec8-kube-api-access-xnghp\") pod \"de559ee7-c4ba-4ebc-93f4-567083524ec8\" (UID: \"de559ee7-c4ba-4ebc-93f4-567083524ec8\") " Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.537732 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de559ee7-c4ba-4ebc-93f4-567083524ec8-catalog-content\") pod \"de559ee7-c4ba-4ebc-93f4-567083524ec8\" (UID: \"de559ee7-c4ba-4ebc-93f4-567083524ec8\") " Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.538924 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de559ee7-c4ba-4ebc-93f4-567083524ec8-utilities\") pod \"de559ee7-c4ba-4ebc-93f4-567083524ec8\" (UID: \"de559ee7-c4ba-4ebc-93f4-567083524ec8\") " Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.539581 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de559ee7-c4ba-4ebc-93f4-567083524ec8-utilities" (OuterVolumeSpecName: "utilities") pod "de559ee7-c4ba-4ebc-93f4-567083524ec8" (UID: "de559ee7-c4ba-4ebc-93f4-567083524ec8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.541730 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de559ee7-c4ba-4ebc-93f4-567083524ec8-kube-api-access-xnghp" (OuterVolumeSpecName: "kube-api-access-xnghp") pod "de559ee7-c4ba-4ebc-93f4-567083524ec8" (UID: "de559ee7-c4ba-4ebc-93f4-567083524ec8"). InnerVolumeSpecName "kube-api-access-xnghp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.636583 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de559ee7-c4ba-4ebc-93f4-567083524ec8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de559ee7-c4ba-4ebc-93f4-567083524ec8" (UID: "de559ee7-c4ba-4ebc-93f4-567083524ec8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.640272 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de559ee7-c4ba-4ebc-93f4-567083524ec8-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.640297 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnghp\" (UniqueName: \"kubernetes.io/projected/de559ee7-c4ba-4ebc-93f4-567083524ec8-kube-api-access-xnghp\") on node \"crc\" DevicePath \"\"" Dec 02 01:49:58 crc kubenswrapper[4884]: I1202 01:49:58.640309 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de559ee7-c4ba-4ebc-93f4-567083524ec8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:49:59 crc kubenswrapper[4884]: I1202 01:49:59.185853 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2h2q5" event={"ID":"de559ee7-c4ba-4ebc-93f4-567083524ec8","Type":"ContainerDied","Data":"205eb4a1f678c60394c6f69957e8b2de6f50b2903fc2f070a523db0569622d3c"} Dec 02 01:49:59 crc kubenswrapper[4884]: I1202 01:49:59.185904 4884 scope.go:117] "RemoveContainer" containerID="34262cdc7dd5ac2db16f9a0c35d6ec43c80d74286e9aca46b28fbc7e58c7e8c7" Dec 02 01:49:59 crc kubenswrapper[4884]: I1202 01:49:59.186012 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2h2q5" Dec 02 01:49:59 crc kubenswrapper[4884]: I1202 01:49:59.198173 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-746c6b4b7b-nzgxl" event={"ID":"291657a1-ad1d-4b85-8717-70a240a7f8e0","Type":"ContainerStarted","Data":"bee9d2ac7724d4c2c45f63df3c8aa51c43e1dbf632ec962937bd69a9abbfc134"} Dec 02 01:49:59 crc kubenswrapper[4884]: I1202 01:49:59.219909 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-746c6b4b7b-nzgxl" podStartSLOduration=2.219886629 podStartE2EDuration="2.219886629s" podCreationTimestamp="2025-12-02 01:49:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:49:59.214182131 +0000 UTC m=+815.890019055" watchObservedRunningTime="2025-12-02 01:49:59.219886629 +0000 UTC m=+815.895723523" Dec 02 01:49:59 crc kubenswrapper[4884]: I1202 01:49:59.237092 4884 scope.go:117] "RemoveContainer" containerID="b7a84f24899b67a103b60ca7de5ddfa92ffa404ab40cd79710879047a2a4a2ac" Dec 02 01:49:59 crc kubenswrapper[4884]: I1202 01:49:59.238441 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2h2q5"] Dec 02 01:49:59 crc kubenswrapper[4884]: I1202 01:49:59.242980 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2h2q5"] Dec 02 01:49:59 crc kubenswrapper[4884]: I1202 01:49:59.254934 4884 scope.go:117] "RemoveContainer" containerID="86cbd784937e15f6b908bddb8fca14161498185105be087d8bc022f9c8b03799" Dec 02 01:49:59 crc kubenswrapper[4884]: I1202 01:49:59.626490 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de559ee7-c4ba-4ebc-93f4-567083524ec8" path="/var/lib/kubelet/pods/de559ee7-c4ba-4ebc-93f4-567083524ec8/volumes" Dec 02 01:50:02 crc kubenswrapper[4884]: I1202 01:50:02.223407 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wmss4" event={"ID":"4a1eae32-edb9-42c6-9a22-62012c26bf23","Type":"ContainerStarted","Data":"5fbe0cb7ece490941adb0ebfe798d2440c6dcabca30cbda0907daf995dbf6561"} Dec 02 01:50:02 crc kubenswrapper[4884]: I1202 01:50:02.225063 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5" event={"ID":"c8a2aedd-caa8-4a87-9429-9443ef9c0fe4","Type":"ContainerStarted","Data":"84424ff26bbb90ea45a260e5fd29544079e0d085fc88c626ae50222cacaf7920"} Dec 02 01:50:02 crc kubenswrapper[4884]: I1202 01:50:02.226854 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-d5d28" event={"ID":"51729f65-2f0f-4a8a-bf7d-1b91f706aa82","Type":"ContainerStarted","Data":"dff32957ecc7a61d25c09091eacaaa892f1f05aa9ff8f108d670f49224e6e8db"} Dec 02 01:50:02 crc kubenswrapper[4884]: I1202 01:50:02.227018 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:50:02 crc kubenswrapper[4884]: I1202 01:50:02.228394 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq" event={"ID":"c783fcfc-44e0-4a67-bc02-f4d0f45f10a9","Type":"ContainerStarted","Data":"e2981c9af808921395a17392f0ffcff4fa9a248aa3fb5729c8eceab660138ad0"} Dec 02 01:50:02 crc kubenswrapper[4884]: I1202 01:50:02.228841 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq" Dec 02 01:50:02 crc kubenswrapper[4884]: I1202 01:50:02.252255 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-6skt5" podStartSLOduration=1.319973077 podStartE2EDuration="5.25223444s" podCreationTimestamp="2025-12-02 01:49:57 +0000 UTC" firstStartedPulling="2025-12-02 01:49:57.772453126 +0000 UTC m=+814.448290010" lastFinishedPulling="2025-12-02 01:50:01.704714489 +0000 UTC m=+818.380551373" observedRunningTime="2025-12-02 01:50:02.250210361 +0000 UTC m=+818.926047285" watchObservedRunningTime="2025-12-02 01:50:02.25223444 +0000 UTC m=+818.928071324" Dec 02 01:50:02 crc kubenswrapper[4884]: I1202 01:50:02.284702 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq" podStartSLOduration=1.501901556 podStartE2EDuration="5.284681352s" podCreationTimestamp="2025-12-02 01:49:57 +0000 UTC" firstStartedPulling="2025-12-02 01:49:57.967021011 +0000 UTC m=+814.642857895" lastFinishedPulling="2025-12-02 01:50:01.749800797 +0000 UTC m=+818.425637691" observedRunningTime="2025-12-02 01:50:02.281917966 +0000 UTC m=+818.957754850" watchObservedRunningTime="2025-12-02 01:50:02.284681352 +0000 UTC m=+818.960518246" Dec 02 01:50:03 crc kubenswrapper[4884]: I1202 01:50:03.647471 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-d5d28" podStartSLOduration=2.444522179 podStartE2EDuration="6.647444262s" podCreationTimestamp="2025-12-02 01:49:57 +0000 UTC" firstStartedPulling="2025-12-02 01:49:57.501208672 +0000 UTC m=+814.177045556" lastFinishedPulling="2025-12-02 01:50:01.704130755 +0000 UTC m=+818.379967639" observedRunningTime="2025-12-02 01:50:02.306790656 +0000 UTC m=+818.982627560" watchObservedRunningTime="2025-12-02 01:50:03.647444262 +0000 UTC m=+820.323281156" Dec 02 01:50:06 crc kubenswrapper[4884]: I1202 01:50:06.273000 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wmss4" event={"ID":"4a1eae32-edb9-42c6-9a22-62012c26bf23","Type":"ContainerStarted","Data":"0f88708ed618da1a83951f3884ca5c7ac410113c1433fe5d57f551f13e7ee8d0"} Dec 02 01:50:06 crc kubenswrapper[4884]: I1202 01:50:06.306833 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-wmss4" podStartSLOduration=1.96752091 podStartE2EDuration="9.306810604s" podCreationTimestamp="2025-12-02 01:49:57 +0000 UTC" firstStartedPulling="2025-12-02 01:49:57.882363648 +0000 UTC m=+814.558200532" lastFinishedPulling="2025-12-02 01:50:05.221653342 +0000 UTC m=+821.897490226" observedRunningTime="2025-12-02 01:50:06.298520154 +0000 UTC m=+822.974357118" watchObservedRunningTime="2025-12-02 01:50:06.306810604 +0000 UTC m=+822.982647498" Dec 02 01:50:07 crc kubenswrapper[4884]: I1202 01:50:07.510119 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-d5d28" Dec 02 01:50:07 crc kubenswrapper[4884]: I1202 01:50:07.882388 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:50:07 crc kubenswrapper[4884]: I1202 01:50:07.882715 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:50:07 crc kubenswrapper[4884]: I1202 01:50:07.890969 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:50:08 crc kubenswrapper[4884]: I1202 01:50:08.295166 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-746c6b4b7b-nzgxl" Dec 02 01:50:08 crc kubenswrapper[4884]: I1202 01:50:08.365334 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-mh86q"] Dec 02 01:50:16 crc kubenswrapper[4884]: I1202 01:50:16.971935 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:50:16 crc kubenswrapper[4884]: I1202 01:50:16.972641 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:50:16 crc kubenswrapper[4884]: I1202 01:50:16.972707 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:50:16 crc kubenswrapper[4884]: I1202 01:50:16.973589 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ec6281f97479113b2e5dc03c077596d5c0e07b1ee23616b661b78bdd0a2d70c8"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 01:50:16 crc kubenswrapper[4884]: I1202 01:50:16.973686 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://ec6281f97479113b2e5dc03c077596d5c0e07b1ee23616b661b78bdd0a2d70c8" gracePeriod=600 Dec 02 01:50:17 crc kubenswrapper[4884]: I1202 01:50:17.362225 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="ec6281f97479113b2e5dc03c077596d5c0e07b1ee23616b661b78bdd0a2d70c8" exitCode=0 Dec 02 01:50:17 crc kubenswrapper[4884]: I1202 01:50:17.362338 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"ec6281f97479113b2e5dc03c077596d5c0e07b1ee23616b661b78bdd0a2d70c8"} Dec 02 01:50:17 crc kubenswrapper[4884]: I1202 01:50:17.362800 4884 scope.go:117] "RemoveContainer" containerID="7a58823165e308a2f2aec8a5502b8d1ea05368d51c2ac26ca666638cb48eebff" Dec 02 01:50:17 crc kubenswrapper[4884]: I1202 01:50:17.750492 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-2mzbq" Dec 02 01:50:18 crc kubenswrapper[4884]: I1202 01:50:18.374888 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"7b403650e63fc1ca4927612983b11a4a56c5610f07757026da030e5a5f4be8dc"} Dec 02 01:50:33 crc kubenswrapper[4884]: I1202 01:50:33.411524 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-mh86q" podUID="d3232fbf-e182-4e1b-a27d-ae2cf61b40b7" containerName="console" containerID="cri-o://fe026586176141e5cba9b7b8025e4f8c7038e36d78f7d10f416b794189287a0f" gracePeriod=15 Dec 02 01:50:33 crc kubenswrapper[4884]: I1202 01:50:33.857073 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-mh86q_d3232fbf-e182-4e1b-a27d-ae2cf61b40b7/console/0.log" Dec 02 01:50:33 crc kubenswrapper[4884]: I1202 01:50:33.857408 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.016605 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-oauth-serving-cert\") pod \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.016678 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-trusted-ca-bundle\") pod \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.016740 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-oauth-config\") pod \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.016833 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96x9n\" (UniqueName: \"kubernetes.io/projected/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-kube-api-access-96x9n\") pod \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.016911 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-config\") pod \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.016946 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-service-ca\") pod \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.016982 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-serving-cert\") pod \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\" (UID: \"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7\") " Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.018072 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "d3232fbf-e182-4e1b-a27d-ae2cf61b40b7" (UID: "d3232fbf-e182-4e1b-a27d-ae2cf61b40b7"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.018124 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "d3232fbf-e182-4e1b-a27d-ae2cf61b40b7" (UID: "d3232fbf-e182-4e1b-a27d-ae2cf61b40b7"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.018110 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-service-ca" (OuterVolumeSpecName: "service-ca") pod "d3232fbf-e182-4e1b-a27d-ae2cf61b40b7" (UID: "d3232fbf-e182-4e1b-a27d-ae2cf61b40b7"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.018168 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-config" (OuterVolumeSpecName: "console-config") pod "d3232fbf-e182-4e1b-a27d-ae2cf61b40b7" (UID: "d3232fbf-e182-4e1b-a27d-ae2cf61b40b7"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.026381 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "d3232fbf-e182-4e1b-a27d-ae2cf61b40b7" (UID: "d3232fbf-e182-4e1b-a27d-ae2cf61b40b7"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.026489 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-kube-api-access-96x9n" (OuterVolumeSpecName: "kube-api-access-96x9n") pod "d3232fbf-e182-4e1b-a27d-ae2cf61b40b7" (UID: "d3232fbf-e182-4e1b-a27d-ae2cf61b40b7"). InnerVolumeSpecName "kube-api-access-96x9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.027342 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "d3232fbf-e182-4e1b-a27d-ae2cf61b40b7" (UID: "d3232fbf-e182-4e1b-a27d-ae2cf61b40b7"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.118408 4884 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.118461 4884 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-service-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.118483 4884 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.118503 4884 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.118522 4884 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.118539 4884 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.118558 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96x9n\" (UniqueName: \"kubernetes.io/projected/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7-kube-api-access-96x9n\") on node \"crc\" DevicePath \"\"" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.500105 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-mh86q_d3232fbf-e182-4e1b-a27d-ae2cf61b40b7/console/0.log" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.500192 4884 generic.go:334] "Generic (PLEG): container finished" podID="d3232fbf-e182-4e1b-a27d-ae2cf61b40b7" containerID="fe026586176141e5cba9b7b8025e4f8c7038e36d78f7d10f416b794189287a0f" exitCode=2 Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.500237 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mh86q" event={"ID":"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7","Type":"ContainerDied","Data":"fe026586176141e5cba9b7b8025e4f8c7038e36d78f7d10f416b794189287a0f"} Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.500279 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mh86q" event={"ID":"d3232fbf-e182-4e1b-a27d-ae2cf61b40b7","Type":"ContainerDied","Data":"82bd2eecf46017729aa2de6ec2c9d8acb3f1f8c4140f498ecab0b728cc87e427"} Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.500309 4884 scope.go:117] "RemoveContainer" containerID="fe026586176141e5cba9b7b8025e4f8c7038e36d78f7d10f416b794189287a0f" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.500317 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mh86q" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.534999 4884 scope.go:117] "RemoveContainer" containerID="fe026586176141e5cba9b7b8025e4f8c7038e36d78f7d10f416b794189287a0f" Dec 02 01:50:34 crc kubenswrapper[4884]: E1202 01:50:34.536325 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe026586176141e5cba9b7b8025e4f8c7038e36d78f7d10f416b794189287a0f\": container with ID starting with fe026586176141e5cba9b7b8025e4f8c7038e36d78f7d10f416b794189287a0f not found: ID does not exist" containerID="fe026586176141e5cba9b7b8025e4f8c7038e36d78f7d10f416b794189287a0f" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.536459 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe026586176141e5cba9b7b8025e4f8c7038e36d78f7d10f416b794189287a0f"} err="failed to get container status \"fe026586176141e5cba9b7b8025e4f8c7038e36d78f7d10f416b794189287a0f\": rpc error: code = NotFound desc = could not find container \"fe026586176141e5cba9b7b8025e4f8c7038e36d78f7d10f416b794189287a0f\": container with ID starting with fe026586176141e5cba9b7b8025e4f8c7038e36d78f7d10f416b794189287a0f not found: ID does not exist" Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.551153 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-mh86q"] Dec 02 01:50:34 crc kubenswrapper[4884]: I1202 01:50:34.556683 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-mh86q"] Dec 02 01:50:35 crc kubenswrapper[4884]: I1202 01:50:35.626520 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3232fbf-e182-4e1b-a27d-ae2cf61b40b7" path="/var/lib/kubelet/pods/d3232fbf-e182-4e1b-a27d-ae2cf61b40b7/volumes" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.008529 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t"] Dec 02 01:50:36 crc kubenswrapper[4884]: E1202 01:50:36.008790 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de559ee7-c4ba-4ebc-93f4-567083524ec8" containerName="extract-utilities" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.008802 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="de559ee7-c4ba-4ebc-93f4-567083524ec8" containerName="extract-utilities" Dec 02 01:50:36 crc kubenswrapper[4884]: E1202 01:50:36.008812 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3232fbf-e182-4e1b-a27d-ae2cf61b40b7" containerName="console" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.008817 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3232fbf-e182-4e1b-a27d-ae2cf61b40b7" containerName="console" Dec 02 01:50:36 crc kubenswrapper[4884]: E1202 01:50:36.008831 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de559ee7-c4ba-4ebc-93f4-567083524ec8" containerName="registry-server" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.008837 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="de559ee7-c4ba-4ebc-93f4-567083524ec8" containerName="registry-server" Dec 02 01:50:36 crc kubenswrapper[4884]: E1202 01:50:36.008845 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de559ee7-c4ba-4ebc-93f4-567083524ec8" containerName="extract-content" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.008851 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="de559ee7-c4ba-4ebc-93f4-567083524ec8" containerName="extract-content" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.008941 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="de559ee7-c4ba-4ebc-93f4-567083524ec8" containerName="registry-server" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.008955 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3232fbf-e182-4e1b-a27d-ae2cf61b40b7" containerName="console" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.009688 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.012697 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.018078 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t"] Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.147208 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2019f363-8aa0-4167-8fdc-ebf932fd07ed-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t\" (UID: \"2019f363-8aa0-4167-8fdc-ebf932fd07ed\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.147284 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2019f363-8aa0-4167-8fdc-ebf932fd07ed-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t\" (UID: \"2019f363-8aa0-4167-8fdc-ebf932fd07ed\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.147329 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhn22\" (UniqueName: \"kubernetes.io/projected/2019f363-8aa0-4167-8fdc-ebf932fd07ed-kube-api-access-vhn22\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t\" (UID: \"2019f363-8aa0-4167-8fdc-ebf932fd07ed\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.248735 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2019f363-8aa0-4167-8fdc-ebf932fd07ed-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t\" (UID: \"2019f363-8aa0-4167-8fdc-ebf932fd07ed\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.248855 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2019f363-8aa0-4167-8fdc-ebf932fd07ed-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t\" (UID: \"2019f363-8aa0-4167-8fdc-ebf932fd07ed\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.248887 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhn22\" (UniqueName: \"kubernetes.io/projected/2019f363-8aa0-4167-8fdc-ebf932fd07ed-kube-api-access-vhn22\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t\" (UID: \"2019f363-8aa0-4167-8fdc-ebf932fd07ed\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.249381 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2019f363-8aa0-4167-8fdc-ebf932fd07ed-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t\" (UID: \"2019f363-8aa0-4167-8fdc-ebf932fd07ed\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.249795 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2019f363-8aa0-4167-8fdc-ebf932fd07ed-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t\" (UID: \"2019f363-8aa0-4167-8fdc-ebf932fd07ed\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.267413 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhn22\" (UniqueName: \"kubernetes.io/projected/2019f363-8aa0-4167-8fdc-ebf932fd07ed-kube-api-access-vhn22\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t\" (UID: \"2019f363-8aa0-4167-8fdc-ebf932fd07ed\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.325007 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" Dec 02 01:50:36 crc kubenswrapper[4884]: I1202 01:50:36.544795 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t"] Dec 02 01:50:37 crc kubenswrapper[4884]: I1202 01:50:37.548103 4884 generic.go:334] "Generic (PLEG): container finished" podID="2019f363-8aa0-4167-8fdc-ebf932fd07ed" containerID="0b4cf039675e80474ace48ca5c880583ab2160f7685500949e0559de06c4db76" exitCode=0 Dec 02 01:50:37 crc kubenswrapper[4884]: I1202 01:50:37.548225 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" event={"ID":"2019f363-8aa0-4167-8fdc-ebf932fd07ed","Type":"ContainerDied","Data":"0b4cf039675e80474ace48ca5c880583ab2160f7685500949e0559de06c4db76"} Dec 02 01:50:37 crc kubenswrapper[4884]: I1202 01:50:37.548851 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" event={"ID":"2019f363-8aa0-4167-8fdc-ebf932fd07ed","Type":"ContainerStarted","Data":"169c6e176a55366afbcb9d04ba53dfe6e98d38701bc1e139fb1103b97453d8da"} Dec 02 01:50:39 crc kubenswrapper[4884]: I1202 01:50:39.566090 4884 generic.go:334] "Generic (PLEG): container finished" podID="2019f363-8aa0-4167-8fdc-ebf932fd07ed" containerID="7503b38e3907505f876f1b66d344798da14456c1ae7162058aef6bb2921b5e17" exitCode=0 Dec 02 01:50:39 crc kubenswrapper[4884]: I1202 01:50:39.566190 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" event={"ID":"2019f363-8aa0-4167-8fdc-ebf932fd07ed","Type":"ContainerDied","Data":"7503b38e3907505f876f1b66d344798da14456c1ae7162058aef6bb2921b5e17"} Dec 02 01:50:40 crc kubenswrapper[4884]: I1202 01:50:40.577713 4884 generic.go:334] "Generic (PLEG): container finished" podID="2019f363-8aa0-4167-8fdc-ebf932fd07ed" containerID="a2c02d462b20fbefce80efc7c75fb4e33303f9b98071b899271d023423cfebfb" exitCode=0 Dec 02 01:50:40 crc kubenswrapper[4884]: I1202 01:50:40.578616 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" event={"ID":"2019f363-8aa0-4167-8fdc-ebf932fd07ed","Type":"ContainerDied","Data":"a2c02d462b20fbefce80efc7c75fb4e33303f9b98071b899271d023423cfebfb"} Dec 02 01:50:41 crc kubenswrapper[4884]: I1202 01:50:41.930635 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" Dec 02 01:50:41 crc kubenswrapper[4884]: I1202 01:50:41.939782 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2019f363-8aa0-4167-8fdc-ebf932fd07ed-bundle\") pod \"2019f363-8aa0-4167-8fdc-ebf932fd07ed\" (UID: \"2019f363-8aa0-4167-8fdc-ebf932fd07ed\") " Dec 02 01:50:41 crc kubenswrapper[4884]: I1202 01:50:41.939819 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhn22\" (UniqueName: \"kubernetes.io/projected/2019f363-8aa0-4167-8fdc-ebf932fd07ed-kube-api-access-vhn22\") pod \"2019f363-8aa0-4167-8fdc-ebf932fd07ed\" (UID: \"2019f363-8aa0-4167-8fdc-ebf932fd07ed\") " Dec 02 01:50:41 crc kubenswrapper[4884]: I1202 01:50:41.939912 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2019f363-8aa0-4167-8fdc-ebf932fd07ed-util\") pod \"2019f363-8aa0-4167-8fdc-ebf932fd07ed\" (UID: \"2019f363-8aa0-4167-8fdc-ebf932fd07ed\") " Dec 02 01:50:41 crc kubenswrapper[4884]: I1202 01:50:41.940987 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2019f363-8aa0-4167-8fdc-ebf932fd07ed-bundle" (OuterVolumeSpecName: "bundle") pod "2019f363-8aa0-4167-8fdc-ebf932fd07ed" (UID: "2019f363-8aa0-4167-8fdc-ebf932fd07ed"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:50:41 crc kubenswrapper[4884]: I1202 01:50:41.949198 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2019f363-8aa0-4167-8fdc-ebf932fd07ed-kube-api-access-vhn22" (OuterVolumeSpecName: "kube-api-access-vhn22") pod "2019f363-8aa0-4167-8fdc-ebf932fd07ed" (UID: "2019f363-8aa0-4167-8fdc-ebf932fd07ed"). InnerVolumeSpecName "kube-api-access-vhn22". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:50:41 crc kubenswrapper[4884]: I1202 01:50:41.963575 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2019f363-8aa0-4167-8fdc-ebf932fd07ed-util" (OuterVolumeSpecName: "util") pod "2019f363-8aa0-4167-8fdc-ebf932fd07ed" (UID: "2019f363-8aa0-4167-8fdc-ebf932fd07ed"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:50:42 crc kubenswrapper[4884]: I1202 01:50:42.041376 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhn22\" (UniqueName: \"kubernetes.io/projected/2019f363-8aa0-4167-8fdc-ebf932fd07ed-kube-api-access-vhn22\") on node \"crc\" DevicePath \"\"" Dec 02 01:50:42 crc kubenswrapper[4884]: I1202 01:50:42.041420 4884 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2019f363-8aa0-4167-8fdc-ebf932fd07ed-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:50:42 crc kubenswrapper[4884]: I1202 01:50:42.041440 4884 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2019f363-8aa0-4167-8fdc-ebf932fd07ed-util\") on node \"crc\" DevicePath \"\"" Dec 02 01:50:42 crc kubenswrapper[4884]: I1202 01:50:42.597062 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" event={"ID":"2019f363-8aa0-4167-8fdc-ebf932fd07ed","Type":"ContainerDied","Data":"169c6e176a55366afbcb9d04ba53dfe6e98d38701bc1e139fb1103b97453d8da"} Dec 02 01:50:42 crc kubenswrapper[4884]: I1202 01:50:42.597119 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="169c6e176a55366afbcb9d04ba53dfe6e98d38701bc1e139fb1103b97453d8da" Dec 02 01:50:42 crc kubenswrapper[4884]: I1202 01:50:42.597278 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t" Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.745735 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nbx59"] Dec 02 01:50:43 crc kubenswrapper[4884]: E1202 01:50:43.746461 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2019f363-8aa0-4167-8fdc-ebf932fd07ed" containerName="util" Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.746480 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2019f363-8aa0-4167-8fdc-ebf932fd07ed" containerName="util" Dec 02 01:50:43 crc kubenswrapper[4884]: E1202 01:50:43.746500 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2019f363-8aa0-4167-8fdc-ebf932fd07ed" containerName="pull" Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.746514 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2019f363-8aa0-4167-8fdc-ebf932fd07ed" containerName="pull" Dec 02 01:50:43 crc kubenswrapper[4884]: E1202 01:50:43.746557 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2019f363-8aa0-4167-8fdc-ebf932fd07ed" containerName="extract" Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.746569 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2019f363-8aa0-4167-8fdc-ebf932fd07ed" containerName="extract" Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.746769 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="2019f363-8aa0-4167-8fdc-ebf932fd07ed" containerName="extract" Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.748075 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.758428 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nbx59"] Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.789942 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmq8l\" (UniqueName: \"kubernetes.io/projected/56293e4c-37a6-4660-99d5-836022aca7dd-kube-api-access-zmq8l\") pod \"community-operators-nbx59\" (UID: \"56293e4c-37a6-4660-99d5-836022aca7dd\") " pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.790001 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56293e4c-37a6-4660-99d5-836022aca7dd-utilities\") pod \"community-operators-nbx59\" (UID: \"56293e4c-37a6-4660-99d5-836022aca7dd\") " pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.790113 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56293e4c-37a6-4660-99d5-836022aca7dd-catalog-content\") pod \"community-operators-nbx59\" (UID: \"56293e4c-37a6-4660-99d5-836022aca7dd\") " pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.891164 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmq8l\" (UniqueName: \"kubernetes.io/projected/56293e4c-37a6-4660-99d5-836022aca7dd-kube-api-access-zmq8l\") pod \"community-operators-nbx59\" (UID: \"56293e4c-37a6-4660-99d5-836022aca7dd\") " pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.891628 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56293e4c-37a6-4660-99d5-836022aca7dd-utilities\") pod \"community-operators-nbx59\" (UID: \"56293e4c-37a6-4660-99d5-836022aca7dd\") " pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.891949 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56293e4c-37a6-4660-99d5-836022aca7dd-catalog-content\") pod \"community-operators-nbx59\" (UID: \"56293e4c-37a6-4660-99d5-836022aca7dd\") " pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.892148 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56293e4c-37a6-4660-99d5-836022aca7dd-utilities\") pod \"community-operators-nbx59\" (UID: \"56293e4c-37a6-4660-99d5-836022aca7dd\") " pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.892518 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56293e4c-37a6-4660-99d5-836022aca7dd-catalog-content\") pod \"community-operators-nbx59\" (UID: \"56293e4c-37a6-4660-99d5-836022aca7dd\") " pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:43 crc kubenswrapper[4884]: I1202 01:50:43.924817 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmq8l\" (UniqueName: \"kubernetes.io/projected/56293e4c-37a6-4660-99d5-836022aca7dd-kube-api-access-zmq8l\") pod \"community-operators-nbx59\" (UID: \"56293e4c-37a6-4660-99d5-836022aca7dd\") " pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:44 crc kubenswrapper[4884]: I1202 01:50:44.075025 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:44 crc kubenswrapper[4884]: I1202 01:50:44.366019 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nbx59"] Dec 02 01:50:44 crc kubenswrapper[4884]: I1202 01:50:44.613017 4884 generic.go:334] "Generic (PLEG): container finished" podID="56293e4c-37a6-4660-99d5-836022aca7dd" containerID="8adde5661b7431db2b6d2d8e56b8a6154836ba1777db3a3cc0c853b4c09e3183" exitCode=0 Dec 02 01:50:44 crc kubenswrapper[4884]: I1202 01:50:44.613060 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbx59" event={"ID":"56293e4c-37a6-4660-99d5-836022aca7dd","Type":"ContainerDied","Data":"8adde5661b7431db2b6d2d8e56b8a6154836ba1777db3a3cc0c853b4c09e3183"} Dec 02 01:50:44 crc kubenswrapper[4884]: I1202 01:50:44.613084 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbx59" event={"ID":"56293e4c-37a6-4660-99d5-836022aca7dd","Type":"ContainerStarted","Data":"9d6dbfe59160aeb607fbf26ffe3f506bfa571d6e05a16e371292e6c5717fdf34"} Dec 02 01:50:45 crc kubenswrapper[4884]: I1202 01:50:45.623567 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbx59" event={"ID":"56293e4c-37a6-4660-99d5-836022aca7dd","Type":"ContainerStarted","Data":"2387b4463a5530cf56d7ec16b38a69297a2af1c81947ea1d290cc3f5f77c78b9"} Dec 02 01:50:45 crc kubenswrapper[4884]: I1202 01:50:45.932113 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g579m"] Dec 02 01:50:45 crc kubenswrapper[4884]: I1202 01:50:45.933706 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:50:45 crc kubenswrapper[4884]: I1202 01:50:45.946391 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g579m"] Dec 02 01:50:46 crc kubenswrapper[4884]: I1202 01:50:46.017289 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f34ae58-5f83-41df-89e4-642fb73ae5c7-catalog-content\") pod \"certified-operators-g579m\" (UID: \"3f34ae58-5f83-41df-89e4-642fb73ae5c7\") " pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:50:46 crc kubenswrapper[4884]: I1202 01:50:46.017421 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f34ae58-5f83-41df-89e4-642fb73ae5c7-utilities\") pod \"certified-operators-g579m\" (UID: \"3f34ae58-5f83-41df-89e4-642fb73ae5c7\") " pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:50:46 crc kubenswrapper[4884]: I1202 01:50:46.017593 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8s2n\" (UniqueName: \"kubernetes.io/projected/3f34ae58-5f83-41df-89e4-642fb73ae5c7-kube-api-access-b8s2n\") pod \"certified-operators-g579m\" (UID: \"3f34ae58-5f83-41df-89e4-642fb73ae5c7\") " pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:50:46 crc kubenswrapper[4884]: I1202 01:50:46.118085 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f34ae58-5f83-41df-89e4-642fb73ae5c7-utilities\") pod \"certified-operators-g579m\" (UID: \"3f34ae58-5f83-41df-89e4-642fb73ae5c7\") " pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:50:46 crc kubenswrapper[4884]: I1202 01:50:46.118168 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8s2n\" (UniqueName: \"kubernetes.io/projected/3f34ae58-5f83-41df-89e4-642fb73ae5c7-kube-api-access-b8s2n\") pod \"certified-operators-g579m\" (UID: \"3f34ae58-5f83-41df-89e4-642fb73ae5c7\") " pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:50:46 crc kubenswrapper[4884]: I1202 01:50:46.118204 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f34ae58-5f83-41df-89e4-642fb73ae5c7-catalog-content\") pod \"certified-operators-g579m\" (UID: \"3f34ae58-5f83-41df-89e4-642fb73ae5c7\") " pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:50:46 crc kubenswrapper[4884]: I1202 01:50:46.118770 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f34ae58-5f83-41df-89e4-642fb73ae5c7-catalog-content\") pod \"certified-operators-g579m\" (UID: \"3f34ae58-5f83-41df-89e4-642fb73ae5c7\") " pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:50:46 crc kubenswrapper[4884]: I1202 01:50:46.118785 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f34ae58-5f83-41df-89e4-642fb73ae5c7-utilities\") pod \"certified-operators-g579m\" (UID: \"3f34ae58-5f83-41df-89e4-642fb73ae5c7\") " pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:50:46 crc kubenswrapper[4884]: I1202 01:50:46.142136 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8s2n\" (UniqueName: \"kubernetes.io/projected/3f34ae58-5f83-41df-89e4-642fb73ae5c7-kube-api-access-b8s2n\") pod \"certified-operators-g579m\" (UID: \"3f34ae58-5f83-41df-89e4-642fb73ae5c7\") " pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:50:46 crc kubenswrapper[4884]: I1202 01:50:46.256316 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:50:46 crc kubenswrapper[4884]: I1202 01:50:46.529236 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g579m"] Dec 02 01:50:46 crc kubenswrapper[4884]: I1202 01:50:46.633267 4884 generic.go:334] "Generic (PLEG): container finished" podID="56293e4c-37a6-4660-99d5-836022aca7dd" containerID="2387b4463a5530cf56d7ec16b38a69297a2af1c81947ea1d290cc3f5f77c78b9" exitCode=0 Dec 02 01:50:46 crc kubenswrapper[4884]: I1202 01:50:46.633356 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbx59" event={"ID":"56293e4c-37a6-4660-99d5-836022aca7dd","Type":"ContainerDied","Data":"2387b4463a5530cf56d7ec16b38a69297a2af1c81947ea1d290cc3f5f77c78b9"} Dec 02 01:50:46 crc kubenswrapper[4884]: I1202 01:50:46.634689 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g579m" event={"ID":"3f34ae58-5f83-41df-89e4-642fb73ae5c7","Type":"ContainerStarted","Data":"24bc6dc3d947f84cdd924ba6c5b0fb6408cd2e4dfdab04290f02ac110d5c79ee"} Dec 02 01:50:47 crc kubenswrapper[4884]: I1202 01:50:47.642931 4884 generic.go:334] "Generic (PLEG): container finished" podID="3f34ae58-5f83-41df-89e4-642fb73ae5c7" containerID="1b1ee52c871c438fd5a9be4c4d7c0e84d103c6041a3c4fcf5fb9539c9035f51a" exitCode=0 Dec 02 01:50:47 crc kubenswrapper[4884]: I1202 01:50:47.643028 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g579m" event={"ID":"3f34ae58-5f83-41df-89e4-642fb73ae5c7","Type":"ContainerDied","Data":"1b1ee52c871c438fd5a9be4c4d7c0e84d103c6041a3c4fcf5fb9539c9035f51a"} Dec 02 01:50:47 crc kubenswrapper[4884]: I1202 01:50:47.647768 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbx59" event={"ID":"56293e4c-37a6-4660-99d5-836022aca7dd","Type":"ContainerStarted","Data":"b702fd23e87ec9413b82dc6fd46cc35bc2eb9a215357d9564fd388c62b181682"} Dec 02 01:50:47 crc kubenswrapper[4884]: I1202 01:50:47.691486 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nbx59" podStartSLOduration=2.228343329 podStartE2EDuration="4.691468442s" podCreationTimestamp="2025-12-02 01:50:43 +0000 UTC" firstStartedPulling="2025-12-02 01:50:44.615159636 +0000 UTC m=+861.290996510" lastFinishedPulling="2025-12-02 01:50:47.078284739 +0000 UTC m=+863.754121623" observedRunningTime="2025-12-02 01:50:47.687911424 +0000 UTC m=+864.363748318" watchObservedRunningTime="2025-12-02 01:50:47.691468442 +0000 UTC m=+864.367305326" Dec 02 01:50:49 crc kubenswrapper[4884]: I1202 01:50:49.662204 4884 generic.go:334] "Generic (PLEG): container finished" podID="3f34ae58-5f83-41df-89e4-642fb73ae5c7" containerID="b8e6915b438136fc84b2b912b53edbcb9b194c6fd985cfda1574ded27c233e58" exitCode=0 Dec 02 01:50:49 crc kubenswrapper[4884]: I1202 01:50:49.662344 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g579m" event={"ID":"3f34ae58-5f83-41df-89e4-642fb73ae5c7","Type":"ContainerDied","Data":"b8e6915b438136fc84b2b912b53edbcb9b194c6fd985cfda1574ded27c233e58"} Dec 02 01:50:50 crc kubenswrapper[4884]: I1202 01:50:50.673014 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g579m" event={"ID":"3f34ae58-5f83-41df-89e4-642fb73ae5c7","Type":"ContainerStarted","Data":"c17c6176b54521ca76d43189aacedc4b639ebecc8caee8613a287497f62a2a01"} Dec 02 01:50:50 crc kubenswrapper[4884]: I1202 01:50:50.695636 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g579m" podStartSLOduration=3.108454485 podStartE2EDuration="5.695617127s" podCreationTimestamp="2025-12-02 01:50:45 +0000 UTC" firstStartedPulling="2025-12-02 01:50:47.64587551 +0000 UTC m=+864.321712434" lastFinishedPulling="2025-12-02 01:50:50.233038182 +0000 UTC m=+866.908875076" observedRunningTime="2025-12-02 01:50:50.692091269 +0000 UTC m=+867.367928163" watchObservedRunningTime="2025-12-02 01:50:50.695617127 +0000 UTC m=+867.371454021" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.349988 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6"] Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.350706 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.352367 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.352639 4884 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.352832 4884 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.353131 4884 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-shd8t" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.353380 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.369921 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6"] Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.490676 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ebe290f2-7fe0-43c7-9485-a7d1789a4460-webhook-cert\") pod \"metallb-operator-controller-manager-864859b84c-8pkv6\" (UID: \"ebe290f2-7fe0-43c7-9485-a7d1789a4460\") " pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.490733 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ebe290f2-7fe0-43c7-9485-a7d1789a4460-apiservice-cert\") pod \"metallb-operator-controller-manager-864859b84c-8pkv6\" (UID: \"ebe290f2-7fe0-43c7-9485-a7d1789a4460\") " pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.490800 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7227\" (UniqueName: \"kubernetes.io/projected/ebe290f2-7fe0-43c7-9485-a7d1789a4460-kube-api-access-v7227\") pod \"metallb-operator-controller-manager-864859b84c-8pkv6\" (UID: \"ebe290f2-7fe0-43c7-9485-a7d1789a4460\") " pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.592704 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ebe290f2-7fe0-43c7-9485-a7d1789a4460-webhook-cert\") pod \"metallb-operator-controller-manager-864859b84c-8pkv6\" (UID: \"ebe290f2-7fe0-43c7-9485-a7d1789a4460\") " pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.592760 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ebe290f2-7fe0-43c7-9485-a7d1789a4460-apiservice-cert\") pod \"metallb-operator-controller-manager-864859b84c-8pkv6\" (UID: \"ebe290f2-7fe0-43c7-9485-a7d1789a4460\") " pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.592811 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7227\" (UniqueName: \"kubernetes.io/projected/ebe290f2-7fe0-43c7-9485-a7d1789a4460-kube-api-access-v7227\") pod \"metallb-operator-controller-manager-864859b84c-8pkv6\" (UID: \"ebe290f2-7fe0-43c7-9485-a7d1789a4460\") " pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.598668 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ebe290f2-7fe0-43c7-9485-a7d1789a4460-apiservice-cert\") pod \"metallb-operator-controller-manager-864859b84c-8pkv6\" (UID: \"ebe290f2-7fe0-43c7-9485-a7d1789a4460\") " pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.603344 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ebe290f2-7fe0-43c7-9485-a7d1789a4460-webhook-cert\") pod \"metallb-operator-controller-manager-864859b84c-8pkv6\" (UID: \"ebe290f2-7fe0-43c7-9485-a7d1789a4460\") " pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.616428 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7227\" (UniqueName: \"kubernetes.io/projected/ebe290f2-7fe0-43c7-9485-a7d1789a4460-kube-api-access-v7227\") pod \"metallb-operator-controller-manager-864859b84c-8pkv6\" (UID: \"ebe290f2-7fe0-43c7-9485-a7d1789a4460\") " pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.663984 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.733484 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx"] Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.734523 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.736924 4884 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-vcfwq" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.741354 4884 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.742076 4884 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.758709 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx"] Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.899084 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rlrn\" (UniqueName: \"kubernetes.io/projected/cf133a27-1ee3-4426-b67a-dd58ce471d16-kube-api-access-9rlrn\") pod \"metallb-operator-webhook-server-6c4c8b8bbb-plfsx\" (UID: \"cf133a27-1ee3-4426-b67a-dd58ce471d16\") " pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.899158 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cf133a27-1ee3-4426-b67a-dd58ce471d16-apiservice-cert\") pod \"metallb-operator-webhook-server-6c4c8b8bbb-plfsx\" (UID: \"cf133a27-1ee3-4426-b67a-dd58ce471d16\") " pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" Dec 02 01:50:51 crc kubenswrapper[4884]: I1202 01:50:51.899212 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cf133a27-1ee3-4426-b67a-dd58ce471d16-webhook-cert\") pod \"metallb-operator-webhook-server-6c4c8b8bbb-plfsx\" (UID: \"cf133a27-1ee3-4426-b67a-dd58ce471d16\") " pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" Dec 02 01:50:52 crc kubenswrapper[4884]: I1202 01:50:52.000355 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cf133a27-1ee3-4426-b67a-dd58ce471d16-webhook-cert\") pod \"metallb-operator-webhook-server-6c4c8b8bbb-plfsx\" (UID: \"cf133a27-1ee3-4426-b67a-dd58ce471d16\") " pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" Dec 02 01:50:52 crc kubenswrapper[4884]: I1202 01:50:52.000401 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rlrn\" (UniqueName: \"kubernetes.io/projected/cf133a27-1ee3-4426-b67a-dd58ce471d16-kube-api-access-9rlrn\") pod \"metallb-operator-webhook-server-6c4c8b8bbb-plfsx\" (UID: \"cf133a27-1ee3-4426-b67a-dd58ce471d16\") " pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" Dec 02 01:50:52 crc kubenswrapper[4884]: I1202 01:50:52.000450 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cf133a27-1ee3-4426-b67a-dd58ce471d16-apiservice-cert\") pod \"metallb-operator-webhook-server-6c4c8b8bbb-plfsx\" (UID: \"cf133a27-1ee3-4426-b67a-dd58ce471d16\") " pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" Dec 02 01:50:52 crc kubenswrapper[4884]: I1202 01:50:52.004839 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cf133a27-1ee3-4426-b67a-dd58ce471d16-webhook-cert\") pod \"metallb-operator-webhook-server-6c4c8b8bbb-plfsx\" (UID: \"cf133a27-1ee3-4426-b67a-dd58ce471d16\") " pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" Dec 02 01:50:52 crc kubenswrapper[4884]: I1202 01:50:52.005232 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cf133a27-1ee3-4426-b67a-dd58ce471d16-apiservice-cert\") pod \"metallb-operator-webhook-server-6c4c8b8bbb-plfsx\" (UID: \"cf133a27-1ee3-4426-b67a-dd58ce471d16\") " pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" Dec 02 01:50:52 crc kubenswrapper[4884]: I1202 01:50:52.022036 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rlrn\" (UniqueName: \"kubernetes.io/projected/cf133a27-1ee3-4426-b67a-dd58ce471d16-kube-api-access-9rlrn\") pod \"metallb-operator-webhook-server-6c4c8b8bbb-plfsx\" (UID: \"cf133a27-1ee3-4426-b67a-dd58ce471d16\") " pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" Dec 02 01:50:52 crc kubenswrapper[4884]: I1202 01:50:52.054367 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" Dec 02 01:50:52 crc kubenswrapper[4884]: I1202 01:50:52.203627 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6"] Dec 02 01:50:52 crc kubenswrapper[4884]: W1202 01:50:52.204717 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebe290f2_7fe0_43c7_9485_a7d1789a4460.slice/crio-e8711c061d5e52b0f4955b0e7e5a6abba380dcaef0cbda05968485413b4e2f95 WatchSource:0}: Error finding container e8711c061d5e52b0f4955b0e7e5a6abba380dcaef0cbda05968485413b4e2f95: Status 404 returned error can't find the container with id e8711c061d5e52b0f4955b0e7e5a6abba380dcaef0cbda05968485413b4e2f95 Dec 02 01:50:52 crc kubenswrapper[4884]: I1202 01:50:52.526384 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx"] Dec 02 01:50:52 crc kubenswrapper[4884]: W1202 01:50:52.540849 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf133a27_1ee3_4426_b67a_dd58ce471d16.slice/crio-31942ab4e78de4edcd45b677ae5ed02c8b7a5dd14f46ffbcf9f8139ed7b06753 WatchSource:0}: Error finding container 31942ab4e78de4edcd45b677ae5ed02c8b7a5dd14f46ffbcf9f8139ed7b06753: Status 404 returned error can't find the container with id 31942ab4e78de4edcd45b677ae5ed02c8b7a5dd14f46ffbcf9f8139ed7b06753 Dec 02 01:50:52 crc kubenswrapper[4884]: I1202 01:50:52.687782 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" event={"ID":"cf133a27-1ee3-4426-b67a-dd58ce471d16","Type":"ContainerStarted","Data":"31942ab4e78de4edcd45b677ae5ed02c8b7a5dd14f46ffbcf9f8139ed7b06753"} Dec 02 01:50:52 crc kubenswrapper[4884]: I1202 01:50:52.689115 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" event={"ID":"ebe290f2-7fe0-43c7-9485-a7d1789a4460","Type":"ContainerStarted","Data":"e8711c061d5e52b0f4955b0e7e5a6abba380dcaef0cbda05968485413b4e2f95"} Dec 02 01:50:53 crc kubenswrapper[4884]: I1202 01:50:53.133199 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nhl9r"] Dec 02 01:50:53 crc kubenswrapper[4884]: I1202 01:50:53.134933 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:50:53 crc kubenswrapper[4884]: I1202 01:50:53.157829 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nhl9r"] Dec 02 01:50:53 crc kubenswrapper[4884]: I1202 01:50:53.243676 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/735c74e3-6c9a-4e60-966b-fbf29a7de36d-utilities\") pod \"redhat-marketplace-nhl9r\" (UID: \"735c74e3-6c9a-4e60-966b-fbf29a7de36d\") " pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:50:53 crc kubenswrapper[4884]: I1202 01:50:53.243762 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nnnm\" (UniqueName: \"kubernetes.io/projected/735c74e3-6c9a-4e60-966b-fbf29a7de36d-kube-api-access-6nnnm\") pod \"redhat-marketplace-nhl9r\" (UID: \"735c74e3-6c9a-4e60-966b-fbf29a7de36d\") " pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:50:53 crc kubenswrapper[4884]: I1202 01:50:53.243839 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/735c74e3-6c9a-4e60-966b-fbf29a7de36d-catalog-content\") pod \"redhat-marketplace-nhl9r\" (UID: \"735c74e3-6c9a-4e60-966b-fbf29a7de36d\") " pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:50:53 crc kubenswrapper[4884]: I1202 01:50:53.345483 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/735c74e3-6c9a-4e60-966b-fbf29a7de36d-utilities\") pod \"redhat-marketplace-nhl9r\" (UID: \"735c74e3-6c9a-4e60-966b-fbf29a7de36d\") " pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:50:53 crc kubenswrapper[4884]: I1202 01:50:53.345543 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nnnm\" (UniqueName: \"kubernetes.io/projected/735c74e3-6c9a-4e60-966b-fbf29a7de36d-kube-api-access-6nnnm\") pod \"redhat-marketplace-nhl9r\" (UID: \"735c74e3-6c9a-4e60-966b-fbf29a7de36d\") " pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:50:53 crc kubenswrapper[4884]: I1202 01:50:53.345595 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/735c74e3-6c9a-4e60-966b-fbf29a7de36d-catalog-content\") pod \"redhat-marketplace-nhl9r\" (UID: \"735c74e3-6c9a-4e60-966b-fbf29a7de36d\") " pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:50:53 crc kubenswrapper[4884]: I1202 01:50:53.346080 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/735c74e3-6c9a-4e60-966b-fbf29a7de36d-catalog-content\") pod \"redhat-marketplace-nhl9r\" (UID: \"735c74e3-6c9a-4e60-966b-fbf29a7de36d\") " pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:50:53 crc kubenswrapper[4884]: I1202 01:50:53.346283 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/735c74e3-6c9a-4e60-966b-fbf29a7de36d-utilities\") pod \"redhat-marketplace-nhl9r\" (UID: \"735c74e3-6c9a-4e60-966b-fbf29a7de36d\") " pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:50:53 crc kubenswrapper[4884]: I1202 01:50:53.366649 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nnnm\" (UniqueName: \"kubernetes.io/projected/735c74e3-6c9a-4e60-966b-fbf29a7de36d-kube-api-access-6nnnm\") pod \"redhat-marketplace-nhl9r\" (UID: \"735c74e3-6c9a-4e60-966b-fbf29a7de36d\") " pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:50:53 crc kubenswrapper[4884]: I1202 01:50:53.461386 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:50:53 crc kubenswrapper[4884]: I1202 01:50:53.947141 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nhl9r"] Dec 02 01:50:54 crc kubenswrapper[4884]: I1202 01:50:54.076218 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:54 crc kubenswrapper[4884]: I1202 01:50:54.076488 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:54 crc kubenswrapper[4884]: I1202 01:50:54.134315 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:54 crc kubenswrapper[4884]: I1202 01:50:54.709327 4884 generic.go:334] "Generic (PLEG): container finished" podID="735c74e3-6c9a-4e60-966b-fbf29a7de36d" containerID="56c988dee9dc2b726bf1061aa7d13bf689578f54d489c6b5251e6e74986d0ec3" exitCode=0 Dec 02 01:50:54 crc kubenswrapper[4884]: I1202 01:50:54.710083 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhl9r" event={"ID":"735c74e3-6c9a-4e60-966b-fbf29a7de36d","Type":"ContainerDied","Data":"56c988dee9dc2b726bf1061aa7d13bf689578f54d489c6b5251e6e74986d0ec3"} Dec 02 01:50:54 crc kubenswrapper[4884]: I1202 01:50:54.710111 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhl9r" event={"ID":"735c74e3-6c9a-4e60-966b-fbf29a7de36d","Type":"ContainerStarted","Data":"011067497e5e2a3e64d866434d94735ca04a219119004cf13cddb4f35b0bd5a9"} Dec 02 01:50:54 crc kubenswrapper[4884]: I1202 01:50:54.799634 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:55 crc kubenswrapper[4884]: I1202 01:50:55.930629 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nbx59"] Dec 02 01:50:56 crc kubenswrapper[4884]: I1202 01:50:56.256902 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:50:56 crc kubenswrapper[4884]: I1202 01:50:56.256951 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:50:56 crc kubenswrapper[4884]: I1202 01:50:56.304833 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:50:56 crc kubenswrapper[4884]: I1202 01:50:56.730228 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nbx59" podUID="56293e4c-37a6-4660-99d5-836022aca7dd" containerName="registry-server" containerID="cri-o://b702fd23e87ec9413b82dc6fd46cc35bc2eb9a215357d9564fd388c62b181682" gracePeriod=2 Dec 02 01:50:56 crc kubenswrapper[4884]: I1202 01:50:56.793404 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:50:57 crc kubenswrapper[4884]: I1202 01:50:57.743682 4884 generic.go:334] "Generic (PLEG): container finished" podID="56293e4c-37a6-4660-99d5-836022aca7dd" containerID="b702fd23e87ec9413b82dc6fd46cc35bc2eb9a215357d9564fd388c62b181682" exitCode=0 Dec 02 01:50:57 crc kubenswrapper[4884]: I1202 01:50:57.743879 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbx59" event={"ID":"56293e4c-37a6-4660-99d5-836022aca7dd","Type":"ContainerDied","Data":"b702fd23e87ec9413b82dc6fd46cc35bc2eb9a215357d9564fd388c62b181682"} Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.534446 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.727800 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmq8l\" (UniqueName: \"kubernetes.io/projected/56293e4c-37a6-4660-99d5-836022aca7dd-kube-api-access-zmq8l\") pod \"56293e4c-37a6-4660-99d5-836022aca7dd\" (UID: \"56293e4c-37a6-4660-99d5-836022aca7dd\") " Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.727890 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56293e4c-37a6-4660-99d5-836022aca7dd-utilities\") pod \"56293e4c-37a6-4660-99d5-836022aca7dd\" (UID: \"56293e4c-37a6-4660-99d5-836022aca7dd\") " Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.727966 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56293e4c-37a6-4660-99d5-836022aca7dd-catalog-content\") pod \"56293e4c-37a6-4660-99d5-836022aca7dd\" (UID: \"56293e4c-37a6-4660-99d5-836022aca7dd\") " Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.728848 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56293e4c-37a6-4660-99d5-836022aca7dd-utilities" (OuterVolumeSpecName: "utilities") pod "56293e4c-37a6-4660-99d5-836022aca7dd" (UID: "56293e4c-37a6-4660-99d5-836022aca7dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.733251 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56293e4c-37a6-4660-99d5-836022aca7dd-kube-api-access-zmq8l" (OuterVolumeSpecName: "kube-api-access-zmq8l") pod "56293e4c-37a6-4660-99d5-836022aca7dd" (UID: "56293e4c-37a6-4660-99d5-836022aca7dd"). InnerVolumeSpecName "kube-api-access-zmq8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.751954 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nbx59" event={"ID":"56293e4c-37a6-4660-99d5-836022aca7dd","Type":"ContainerDied","Data":"9d6dbfe59160aeb607fbf26ffe3f506bfa571d6e05a16e371292e6c5717fdf34"} Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.751983 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nbx59" Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.752003 4884 scope.go:117] "RemoveContainer" containerID="b702fd23e87ec9413b82dc6fd46cc35bc2eb9a215357d9564fd388c62b181682" Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.755125 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" event={"ID":"cf133a27-1ee3-4426-b67a-dd58ce471d16","Type":"ContainerStarted","Data":"4f1456c0cd010f024480edaab7488bc95aa8f5106abbca50aff82c9e3bc5f2c5"} Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.755148 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.756894 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" event={"ID":"ebe290f2-7fe0-43c7-9485-a7d1789a4460","Type":"ContainerStarted","Data":"5105ce0bf5c2139db6c35d76a231c9a880248c8c6a76a0ed727a5d90a2fb69d4"} Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.757011 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.758321 4884 generic.go:334] "Generic (PLEG): container finished" podID="735c74e3-6c9a-4e60-966b-fbf29a7de36d" containerID="786be6eb1445bfd38f0b28d23ff84c8157aac6e2c1805fe5fbdc62296aac00ea" exitCode=0 Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.758363 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhl9r" event={"ID":"735c74e3-6c9a-4e60-966b-fbf29a7de36d","Type":"ContainerDied","Data":"786be6eb1445bfd38f0b28d23ff84c8157aac6e2c1805fe5fbdc62296aac00ea"} Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.779602 4884 scope.go:117] "RemoveContainer" containerID="2387b4463a5530cf56d7ec16b38a69297a2af1c81947ea1d290cc3f5f77c78b9" Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.782575 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" podStartSLOduration=2.051608871 podStartE2EDuration="7.782557234s" podCreationTimestamp="2025-12-02 01:50:51 +0000 UTC" firstStartedPulling="2025-12-02 01:50:52.545970687 +0000 UTC m=+869.221807581" lastFinishedPulling="2025-12-02 01:50:58.27691904 +0000 UTC m=+874.952755944" observedRunningTime="2025-12-02 01:50:58.780455241 +0000 UTC m=+875.456292125" watchObservedRunningTime="2025-12-02 01:50:58.782557234 +0000 UTC m=+875.458394118" Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.785199 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56293e4c-37a6-4660-99d5-836022aca7dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56293e4c-37a6-4660-99d5-836022aca7dd" (UID: "56293e4c-37a6-4660-99d5-836022aca7dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.799934 4884 scope.go:117] "RemoveContainer" containerID="8adde5661b7431db2b6d2d8e56b8a6154836ba1777db3a3cc0c853b4c09e3183" Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.822317 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" podStartSLOduration=1.772736206 podStartE2EDuration="7.82229929s" podCreationTimestamp="2025-12-02 01:50:51 +0000 UTC" firstStartedPulling="2025-12-02 01:50:52.209674217 +0000 UTC m=+868.885511101" lastFinishedPulling="2025-12-02 01:50:58.259237291 +0000 UTC m=+874.935074185" observedRunningTime="2025-12-02 01:50:58.813689456 +0000 UTC m=+875.489526360" watchObservedRunningTime="2025-12-02 01:50:58.82229929 +0000 UTC m=+875.498136174" Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.829818 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmq8l\" (UniqueName: \"kubernetes.io/projected/56293e4c-37a6-4660-99d5-836022aca7dd-kube-api-access-zmq8l\") on node \"crc\" DevicePath \"\"" Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.829847 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56293e4c-37a6-4660-99d5-836022aca7dd-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:50:58 crc kubenswrapper[4884]: I1202 01:50:58.829856 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56293e4c-37a6-4660-99d5-836022aca7dd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:50:59 crc kubenswrapper[4884]: I1202 01:50:59.098474 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nbx59"] Dec 02 01:50:59 crc kubenswrapper[4884]: I1202 01:50:59.107852 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nbx59"] Dec 02 01:50:59 crc kubenswrapper[4884]: I1202 01:50:59.623572 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56293e4c-37a6-4660-99d5-836022aca7dd" path="/var/lib/kubelet/pods/56293e4c-37a6-4660-99d5-836022aca7dd/volumes" Dec 02 01:51:00 crc kubenswrapper[4884]: I1202 01:51:00.789539 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhl9r" event={"ID":"735c74e3-6c9a-4e60-966b-fbf29a7de36d","Type":"ContainerStarted","Data":"fee0c45aec3be37ce007c5f47e4c554db2054d55830e5f7021041925f0d6328f"} Dec 02 01:51:00 crc kubenswrapper[4884]: I1202 01:51:00.819401 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nhl9r" podStartSLOduration=2.929194263 podStartE2EDuration="7.819380563s" podCreationTimestamp="2025-12-02 01:50:53 +0000 UTC" firstStartedPulling="2025-12-02 01:50:54.711512011 +0000 UTC m=+871.387348885" lastFinishedPulling="2025-12-02 01:50:59.601698291 +0000 UTC m=+876.277535185" observedRunningTime="2025-12-02 01:51:00.814480881 +0000 UTC m=+877.490317855" watchObservedRunningTime="2025-12-02 01:51:00.819380563 +0000 UTC m=+877.495217457" Dec 02 01:51:02 crc kubenswrapper[4884]: I1202 01:51:02.528120 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g579m"] Dec 02 01:51:02 crc kubenswrapper[4884]: I1202 01:51:02.528545 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g579m" podUID="3f34ae58-5f83-41df-89e4-642fb73ae5c7" containerName="registry-server" containerID="cri-o://c17c6176b54521ca76d43189aacedc4b639ebecc8caee8613a287497f62a2a01" gracePeriod=2 Dec 02 01:51:02 crc kubenswrapper[4884]: I1202 01:51:02.808583 4884 generic.go:334] "Generic (PLEG): container finished" podID="3f34ae58-5f83-41df-89e4-642fb73ae5c7" containerID="c17c6176b54521ca76d43189aacedc4b639ebecc8caee8613a287497f62a2a01" exitCode=0 Dec 02 01:51:02 crc kubenswrapper[4884]: I1202 01:51:02.808635 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g579m" event={"ID":"3f34ae58-5f83-41df-89e4-642fb73ae5c7","Type":"ContainerDied","Data":"c17c6176b54521ca76d43189aacedc4b639ebecc8caee8613a287497f62a2a01"} Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.462437 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.462489 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.473106 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.523719 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.594163 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f34ae58-5f83-41df-89e4-642fb73ae5c7-catalog-content\") pod \"3f34ae58-5f83-41df-89e4-642fb73ae5c7\" (UID: \"3f34ae58-5f83-41df-89e4-642fb73ae5c7\") " Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.594222 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8s2n\" (UniqueName: \"kubernetes.io/projected/3f34ae58-5f83-41df-89e4-642fb73ae5c7-kube-api-access-b8s2n\") pod \"3f34ae58-5f83-41df-89e4-642fb73ae5c7\" (UID: \"3f34ae58-5f83-41df-89e4-642fb73ae5c7\") " Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.594273 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f34ae58-5f83-41df-89e4-642fb73ae5c7-utilities\") pod \"3f34ae58-5f83-41df-89e4-642fb73ae5c7\" (UID: \"3f34ae58-5f83-41df-89e4-642fb73ae5c7\") " Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.596634 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f34ae58-5f83-41df-89e4-642fb73ae5c7-utilities" (OuterVolumeSpecName: "utilities") pod "3f34ae58-5f83-41df-89e4-642fb73ae5c7" (UID: "3f34ae58-5f83-41df-89e4-642fb73ae5c7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.604056 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f34ae58-5f83-41df-89e4-642fb73ae5c7-kube-api-access-b8s2n" (OuterVolumeSpecName: "kube-api-access-b8s2n") pod "3f34ae58-5f83-41df-89e4-642fb73ae5c7" (UID: "3f34ae58-5f83-41df-89e4-642fb73ae5c7"). InnerVolumeSpecName "kube-api-access-b8s2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.646492 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f34ae58-5f83-41df-89e4-642fb73ae5c7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f34ae58-5f83-41df-89e4-642fb73ae5c7" (UID: "3f34ae58-5f83-41df-89e4-642fb73ae5c7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.695697 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f34ae58-5f83-41df-89e4-642fb73ae5c7-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.695789 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f34ae58-5f83-41df-89e4-642fb73ae5c7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.695829 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8s2n\" (UniqueName: \"kubernetes.io/projected/3f34ae58-5f83-41df-89e4-642fb73ae5c7-kube-api-access-b8s2n\") on node \"crc\" DevicePath \"\"" Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.816461 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g579m" event={"ID":"3f34ae58-5f83-41df-89e4-642fb73ae5c7","Type":"ContainerDied","Data":"24bc6dc3d947f84cdd924ba6c5b0fb6408cd2e4dfdab04290f02ac110d5c79ee"} Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.816497 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g579m" Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.816518 4884 scope.go:117] "RemoveContainer" containerID="c17c6176b54521ca76d43189aacedc4b639ebecc8caee8613a287497f62a2a01" Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.848402 4884 scope.go:117] "RemoveContainer" containerID="b8e6915b438136fc84b2b912b53edbcb9b194c6fd985cfda1574ded27c233e58" Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.851801 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g579m"] Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.857284 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g579m"] Dec 02 01:51:03 crc kubenswrapper[4884]: I1202 01:51:03.879736 4884 scope.go:117] "RemoveContainer" containerID="1b1ee52c871c438fd5a9be4c4d7c0e84d103c6041a3c4fcf5fb9539c9035f51a" Dec 02 01:51:05 crc kubenswrapper[4884]: I1202 01:51:05.622085 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f34ae58-5f83-41df-89e4-642fb73ae5c7" path="/var/lib/kubelet/pods/3f34ae58-5f83-41df-89e4-642fb73ae5c7/volumes" Dec 02 01:51:12 crc kubenswrapper[4884]: I1202 01:51:12.061963 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6c4c8b8bbb-plfsx" Dec 02 01:51:13 crc kubenswrapper[4884]: I1202 01:51:13.529511 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:51:13 crc kubenswrapper[4884]: I1202 01:51:13.593127 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nhl9r"] Dec 02 01:51:13 crc kubenswrapper[4884]: I1202 01:51:13.885337 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nhl9r" podUID="735c74e3-6c9a-4e60-966b-fbf29a7de36d" containerName="registry-server" containerID="cri-o://fee0c45aec3be37ce007c5f47e4c554db2054d55830e5f7021041925f0d6328f" gracePeriod=2 Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.829125 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.846776 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/735c74e3-6c9a-4e60-966b-fbf29a7de36d-utilities\") pod \"735c74e3-6c9a-4e60-966b-fbf29a7de36d\" (UID: \"735c74e3-6c9a-4e60-966b-fbf29a7de36d\") " Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.846845 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/735c74e3-6c9a-4e60-966b-fbf29a7de36d-catalog-content\") pod \"735c74e3-6c9a-4e60-966b-fbf29a7de36d\" (UID: \"735c74e3-6c9a-4e60-966b-fbf29a7de36d\") " Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.846911 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nnnm\" (UniqueName: \"kubernetes.io/projected/735c74e3-6c9a-4e60-966b-fbf29a7de36d-kube-api-access-6nnnm\") pod \"735c74e3-6c9a-4e60-966b-fbf29a7de36d\" (UID: \"735c74e3-6c9a-4e60-966b-fbf29a7de36d\") " Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.848249 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/735c74e3-6c9a-4e60-966b-fbf29a7de36d-utilities" (OuterVolumeSpecName: "utilities") pod "735c74e3-6c9a-4e60-966b-fbf29a7de36d" (UID: "735c74e3-6c9a-4e60-966b-fbf29a7de36d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.864899 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/735c74e3-6c9a-4e60-966b-fbf29a7de36d-kube-api-access-6nnnm" (OuterVolumeSpecName: "kube-api-access-6nnnm") pod "735c74e3-6c9a-4e60-966b-fbf29a7de36d" (UID: "735c74e3-6c9a-4e60-966b-fbf29a7de36d"). InnerVolumeSpecName "kube-api-access-6nnnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.902333 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/735c74e3-6c9a-4e60-966b-fbf29a7de36d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "735c74e3-6c9a-4e60-966b-fbf29a7de36d" (UID: "735c74e3-6c9a-4e60-966b-fbf29a7de36d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.905056 4884 generic.go:334] "Generic (PLEG): container finished" podID="735c74e3-6c9a-4e60-966b-fbf29a7de36d" containerID="fee0c45aec3be37ce007c5f47e4c554db2054d55830e5f7021041925f0d6328f" exitCode=0 Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.905105 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhl9r" event={"ID":"735c74e3-6c9a-4e60-966b-fbf29a7de36d","Type":"ContainerDied","Data":"fee0c45aec3be37ce007c5f47e4c554db2054d55830e5f7021041925f0d6328f"} Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.905163 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nhl9r" event={"ID":"735c74e3-6c9a-4e60-966b-fbf29a7de36d","Type":"ContainerDied","Data":"011067497e5e2a3e64d866434d94735ca04a219119004cf13cddb4f35b0bd5a9"} Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.905185 4884 scope.go:117] "RemoveContainer" containerID="fee0c45aec3be37ce007c5f47e4c554db2054d55830e5f7021041925f0d6328f" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.905243 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nhl9r" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.926219 4884 scope.go:117] "RemoveContainer" containerID="786be6eb1445bfd38f0b28d23ff84c8157aac6e2c1805fe5fbdc62296aac00ea" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.939661 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nhl9r"] Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.944809 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nhl9r"] Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.949525 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/735c74e3-6c9a-4e60-966b-fbf29a7de36d-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.949632 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/735c74e3-6c9a-4e60-966b-fbf29a7de36d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.949734 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nnnm\" (UniqueName: \"kubernetes.io/projected/735c74e3-6c9a-4e60-966b-fbf29a7de36d-kube-api-access-6nnnm\") on node \"crc\" DevicePath \"\"" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.955380 4884 scope.go:117] "RemoveContainer" containerID="56c988dee9dc2b726bf1061aa7d13bf689578f54d489c6b5251e6e74986d0ec3" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.979701 4884 scope.go:117] "RemoveContainer" containerID="fee0c45aec3be37ce007c5f47e4c554db2054d55830e5f7021041925f0d6328f" Dec 02 01:51:14 crc kubenswrapper[4884]: E1202 01:51:14.983360 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fee0c45aec3be37ce007c5f47e4c554db2054d55830e5f7021041925f0d6328f\": container with ID starting with fee0c45aec3be37ce007c5f47e4c554db2054d55830e5f7021041925f0d6328f not found: ID does not exist" containerID="fee0c45aec3be37ce007c5f47e4c554db2054d55830e5f7021041925f0d6328f" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.983418 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fee0c45aec3be37ce007c5f47e4c554db2054d55830e5f7021041925f0d6328f"} err="failed to get container status \"fee0c45aec3be37ce007c5f47e4c554db2054d55830e5f7021041925f0d6328f\": rpc error: code = NotFound desc = could not find container \"fee0c45aec3be37ce007c5f47e4c554db2054d55830e5f7021041925f0d6328f\": container with ID starting with fee0c45aec3be37ce007c5f47e4c554db2054d55830e5f7021041925f0d6328f not found: ID does not exist" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.983458 4884 scope.go:117] "RemoveContainer" containerID="786be6eb1445bfd38f0b28d23ff84c8157aac6e2c1805fe5fbdc62296aac00ea" Dec 02 01:51:14 crc kubenswrapper[4884]: E1202 01:51:14.983811 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"786be6eb1445bfd38f0b28d23ff84c8157aac6e2c1805fe5fbdc62296aac00ea\": container with ID starting with 786be6eb1445bfd38f0b28d23ff84c8157aac6e2c1805fe5fbdc62296aac00ea not found: ID does not exist" containerID="786be6eb1445bfd38f0b28d23ff84c8157aac6e2c1805fe5fbdc62296aac00ea" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.983849 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"786be6eb1445bfd38f0b28d23ff84c8157aac6e2c1805fe5fbdc62296aac00ea"} err="failed to get container status \"786be6eb1445bfd38f0b28d23ff84c8157aac6e2c1805fe5fbdc62296aac00ea\": rpc error: code = NotFound desc = could not find container \"786be6eb1445bfd38f0b28d23ff84c8157aac6e2c1805fe5fbdc62296aac00ea\": container with ID starting with 786be6eb1445bfd38f0b28d23ff84c8157aac6e2c1805fe5fbdc62296aac00ea not found: ID does not exist" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.983872 4884 scope.go:117] "RemoveContainer" containerID="56c988dee9dc2b726bf1061aa7d13bf689578f54d489c6b5251e6e74986d0ec3" Dec 02 01:51:14 crc kubenswrapper[4884]: E1202 01:51:14.984073 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56c988dee9dc2b726bf1061aa7d13bf689578f54d489c6b5251e6e74986d0ec3\": container with ID starting with 56c988dee9dc2b726bf1061aa7d13bf689578f54d489c6b5251e6e74986d0ec3 not found: ID does not exist" containerID="56c988dee9dc2b726bf1061aa7d13bf689578f54d489c6b5251e6e74986d0ec3" Dec 02 01:51:14 crc kubenswrapper[4884]: I1202 01:51:14.984095 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56c988dee9dc2b726bf1061aa7d13bf689578f54d489c6b5251e6e74986d0ec3"} err="failed to get container status \"56c988dee9dc2b726bf1061aa7d13bf689578f54d489c6b5251e6e74986d0ec3\": rpc error: code = NotFound desc = could not find container \"56c988dee9dc2b726bf1061aa7d13bf689578f54d489c6b5251e6e74986d0ec3\": container with ID starting with 56c988dee9dc2b726bf1061aa7d13bf689578f54d489c6b5251e6e74986d0ec3 not found: ID does not exist" Dec 02 01:51:15 crc kubenswrapper[4884]: I1202 01:51:15.626986 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="735c74e3-6c9a-4e60-966b-fbf29a7de36d" path="/var/lib/kubelet/pods/735c74e3-6c9a-4e60-966b-fbf29a7de36d/volumes" Dec 02 01:51:31 crc kubenswrapper[4884]: I1202 01:51:31.667251 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-864859b84c-8pkv6" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.594378 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm"] Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.594629 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="735c74e3-6c9a-4e60-966b-fbf29a7de36d" containerName="extract-content" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.594644 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="735c74e3-6c9a-4e60-966b-fbf29a7de36d" containerName="extract-content" Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.594659 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56293e4c-37a6-4660-99d5-836022aca7dd" containerName="extract-utilities" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.594665 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="56293e4c-37a6-4660-99d5-836022aca7dd" containerName="extract-utilities" Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.594680 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56293e4c-37a6-4660-99d5-836022aca7dd" containerName="extract-content" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.594686 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="56293e4c-37a6-4660-99d5-836022aca7dd" containerName="extract-content" Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.594696 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f34ae58-5f83-41df-89e4-642fb73ae5c7" containerName="extract-content" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.594702 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f34ae58-5f83-41df-89e4-642fb73ae5c7" containerName="extract-content" Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.594710 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="735c74e3-6c9a-4e60-966b-fbf29a7de36d" containerName="extract-utilities" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.594717 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="735c74e3-6c9a-4e60-966b-fbf29a7de36d" containerName="extract-utilities" Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.594728 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56293e4c-37a6-4660-99d5-836022aca7dd" containerName="registry-server" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.594734 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="56293e4c-37a6-4660-99d5-836022aca7dd" containerName="registry-server" Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.594755 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f34ae58-5f83-41df-89e4-642fb73ae5c7" containerName="registry-server" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.594761 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f34ae58-5f83-41df-89e4-642fb73ae5c7" containerName="registry-server" Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.594771 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f34ae58-5f83-41df-89e4-642fb73ae5c7" containerName="extract-utilities" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.594776 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f34ae58-5f83-41df-89e4-642fb73ae5c7" containerName="extract-utilities" Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.594784 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="735c74e3-6c9a-4e60-966b-fbf29a7de36d" containerName="registry-server" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.594789 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="735c74e3-6c9a-4e60-966b-fbf29a7de36d" containerName="registry-server" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.594896 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f34ae58-5f83-41df-89e4-642fb73ae5c7" containerName="registry-server" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.594910 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="735c74e3-6c9a-4e60-966b-fbf29a7de36d" containerName="registry-server" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.594923 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="56293e4c-37a6-4660-99d5-836022aca7dd" containerName="registry-server" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.595357 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.597967 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-j7xs4"] Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.599574 4884 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-4r6bf" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.603357 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.605659 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.606113 4884 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.611189 4884 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.613821 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm"] Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.625425 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-frr-startup\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.625479 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-reloader\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.625507 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-metrics-certs\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.625572 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbhwp\" (UniqueName: \"kubernetes.io/projected/7ef4b609-a02d-4050-a5cc-768308bf2dcb-kube-api-access-pbhwp\") pod \"frr-k8s-webhook-server-7fcb986d4-5j9vm\" (UID: \"7ef4b609-a02d-4050-a5cc-768308bf2dcb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.625596 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-frr-sockets\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.625622 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-frr-conf\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.625647 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-metrics\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.625681 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkb6p\" (UniqueName: \"kubernetes.io/projected/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-kube-api-access-wkb6p\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.625714 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7ef4b609-a02d-4050-a5cc-768308bf2dcb-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-5j9vm\" (UID: \"7ef4b609-a02d-4050-a5cc-768308bf2dcb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.704119 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-bxtbh"] Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.705124 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bxtbh" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.707351 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-8wv2n"] Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.708171 4884 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.708311 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.708383 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-8wv2n" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.713043 4884 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-czscd" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.715438 4884 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.717580 4884 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.725554 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-8wv2n"] Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727119 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-metrics\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727149 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-metrics-certs\") pod \"speaker-bxtbh\" (UID: \"f3fc51b3-99fe-4f7f-be90-68acd111b7ca\") " pod="metallb-system/speaker-bxtbh" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727173 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e365201e-b871-4206-8dcd-9b87a9452f83-metrics-certs\") pod \"controller-f8648f98b-8wv2n\" (UID: \"e365201e-b871-4206-8dcd-9b87a9452f83\") " pod="metallb-system/controller-f8648f98b-8wv2n" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727192 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkb6p\" (UniqueName: \"kubernetes.io/projected/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-kube-api-access-wkb6p\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727218 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7ef4b609-a02d-4050-a5cc-768308bf2dcb-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-5j9vm\" (UID: \"7ef4b609-a02d-4050-a5cc-768308bf2dcb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727265 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-frr-startup\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727289 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-reloader\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727303 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e365201e-b871-4206-8dcd-9b87a9452f83-cert\") pod \"controller-f8648f98b-8wv2n\" (UID: \"e365201e-b871-4206-8dcd-9b87a9452f83\") " pod="metallb-system/controller-f8648f98b-8wv2n" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727347 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-metrics-certs\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727370 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv2k5\" (UniqueName: \"kubernetes.io/projected/e365201e-b871-4206-8dcd-9b87a9452f83-kube-api-access-wv2k5\") pod \"controller-f8648f98b-8wv2n\" (UID: \"e365201e-b871-4206-8dcd-9b87a9452f83\") " pod="metallb-system/controller-f8648f98b-8wv2n" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727400 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrw6k\" (UniqueName: \"kubernetes.io/projected/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-kube-api-access-wrw6k\") pod \"speaker-bxtbh\" (UID: \"f3fc51b3-99fe-4f7f-be90-68acd111b7ca\") " pod="metallb-system/speaker-bxtbh" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727415 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-metallb-excludel2\") pod \"speaker-bxtbh\" (UID: \"f3fc51b3-99fe-4f7f-be90-68acd111b7ca\") " pod="metallb-system/speaker-bxtbh" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727435 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbhwp\" (UniqueName: \"kubernetes.io/projected/7ef4b609-a02d-4050-a5cc-768308bf2dcb-kube-api-access-pbhwp\") pod \"frr-k8s-webhook-server-7fcb986d4-5j9vm\" (UID: \"7ef4b609-a02d-4050-a5cc-768308bf2dcb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727453 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-frr-sockets\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727467 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-memberlist\") pod \"speaker-bxtbh\" (UID: \"f3fc51b3-99fe-4f7f-be90-68acd111b7ca\") " pod="metallb-system/speaker-bxtbh" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727486 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-frr-conf\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727814 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-frr-conf\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.727990 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-metrics\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.728258 4884 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.728298 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7ef4b609-a02d-4050-a5cc-768308bf2dcb-cert podName:7ef4b609-a02d-4050-a5cc-768308bf2dcb nodeName:}" failed. No retries permitted until 2025-12-02 01:51:33.228284883 +0000 UTC m=+909.904121767 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7ef4b609-a02d-4050-a5cc-768308bf2dcb-cert") pod "frr-k8s-webhook-server-7fcb986d4-5j9vm" (UID: "7ef4b609-a02d-4050-a5cc-768308bf2dcb") : secret "frr-k8s-webhook-server-cert" not found Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.728618 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-reloader\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.728706 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-frr-sockets\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.729814 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-frr-startup\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.748348 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-metrics-certs\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.758829 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbhwp\" (UniqueName: \"kubernetes.io/projected/7ef4b609-a02d-4050-a5cc-768308bf2dcb-kube-api-access-pbhwp\") pod \"frr-k8s-webhook-server-7fcb986d4-5j9vm\" (UID: \"7ef4b609-a02d-4050-a5cc-768308bf2dcb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.773339 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkb6p\" (UniqueName: \"kubernetes.io/projected/d0e2e152-8ae0-495d-835c-5dca5ef66b6c-kube-api-access-wkb6p\") pod \"frr-k8s-j7xs4\" (UID: \"d0e2e152-8ae0-495d-835c-5dca5ef66b6c\") " pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.834317 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-memberlist\") pod \"speaker-bxtbh\" (UID: \"f3fc51b3-99fe-4f7f-be90-68acd111b7ca\") " pod="metallb-system/speaker-bxtbh" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.834376 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-metrics-certs\") pod \"speaker-bxtbh\" (UID: \"f3fc51b3-99fe-4f7f-be90-68acd111b7ca\") " pod="metallb-system/speaker-bxtbh" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.834399 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e365201e-b871-4206-8dcd-9b87a9452f83-metrics-certs\") pod \"controller-f8648f98b-8wv2n\" (UID: \"e365201e-b871-4206-8dcd-9b87a9452f83\") " pod="metallb-system/controller-f8648f98b-8wv2n" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.834453 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e365201e-b871-4206-8dcd-9b87a9452f83-cert\") pod \"controller-f8648f98b-8wv2n\" (UID: \"e365201e-b871-4206-8dcd-9b87a9452f83\") " pod="metallb-system/controller-f8648f98b-8wv2n" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.834476 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv2k5\" (UniqueName: \"kubernetes.io/projected/e365201e-b871-4206-8dcd-9b87a9452f83-kube-api-access-wv2k5\") pod \"controller-f8648f98b-8wv2n\" (UID: \"e365201e-b871-4206-8dcd-9b87a9452f83\") " pod="metallb-system/controller-f8648f98b-8wv2n" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.834506 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrw6k\" (UniqueName: \"kubernetes.io/projected/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-kube-api-access-wrw6k\") pod \"speaker-bxtbh\" (UID: \"f3fc51b3-99fe-4f7f-be90-68acd111b7ca\") " pod="metallb-system/speaker-bxtbh" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.834523 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-metallb-excludel2\") pod \"speaker-bxtbh\" (UID: \"f3fc51b3-99fe-4f7f-be90-68acd111b7ca\") " pod="metallb-system/speaker-bxtbh" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.835138 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-metallb-excludel2\") pod \"speaker-bxtbh\" (UID: \"f3fc51b3-99fe-4f7f-be90-68acd111b7ca\") " pod="metallb-system/speaker-bxtbh" Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.835215 4884 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.835254 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-memberlist podName:f3fc51b3-99fe-4f7f-be90-68acd111b7ca nodeName:}" failed. No retries permitted until 2025-12-02 01:51:33.335243449 +0000 UTC m=+910.011080333 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-memberlist") pod "speaker-bxtbh" (UID: "f3fc51b3-99fe-4f7f-be90-68acd111b7ca") : secret "metallb-memberlist" not found Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.835454 4884 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.835478 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-metrics-certs podName:f3fc51b3-99fe-4f7f-be90-68acd111b7ca nodeName:}" failed. No retries permitted until 2025-12-02 01:51:33.335471735 +0000 UTC m=+910.011308619 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-metrics-certs") pod "speaker-bxtbh" (UID: "f3fc51b3-99fe-4f7f-be90-68acd111b7ca") : secret "speaker-certs-secret" not found Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.835509 4884 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 02 01:51:32 crc kubenswrapper[4884]: E1202 01:51:32.835526 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e365201e-b871-4206-8dcd-9b87a9452f83-metrics-certs podName:e365201e-b871-4206-8dcd-9b87a9452f83 nodeName:}" failed. No retries permitted until 2025-12-02 01:51:33.335520706 +0000 UTC m=+910.011357590 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e365201e-b871-4206-8dcd-9b87a9452f83-metrics-certs") pod "controller-f8648f98b-8wv2n" (UID: "e365201e-b871-4206-8dcd-9b87a9452f83") : secret "controller-certs-secret" not found Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.839953 4884 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.861379 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv2k5\" (UniqueName: \"kubernetes.io/projected/e365201e-b871-4206-8dcd-9b87a9452f83-kube-api-access-wv2k5\") pod \"controller-f8648f98b-8wv2n\" (UID: \"e365201e-b871-4206-8dcd-9b87a9452f83\") " pod="metallb-system/controller-f8648f98b-8wv2n" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.867579 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e365201e-b871-4206-8dcd-9b87a9452f83-cert\") pod \"controller-f8648f98b-8wv2n\" (UID: \"e365201e-b871-4206-8dcd-9b87a9452f83\") " pod="metallb-system/controller-f8648f98b-8wv2n" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.879241 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrw6k\" (UniqueName: \"kubernetes.io/projected/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-kube-api-access-wrw6k\") pod \"speaker-bxtbh\" (UID: \"f3fc51b3-99fe-4f7f-be90-68acd111b7ca\") " pod="metallb-system/speaker-bxtbh" Dec 02 01:51:32 crc kubenswrapper[4884]: I1202 01:51:32.918658 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:33 crc kubenswrapper[4884]: I1202 01:51:33.239573 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7ef4b609-a02d-4050-a5cc-768308bf2dcb-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-5j9vm\" (UID: \"7ef4b609-a02d-4050-a5cc-768308bf2dcb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm" Dec 02 01:51:33 crc kubenswrapper[4884]: I1202 01:51:33.246315 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7ef4b609-a02d-4050-a5cc-768308bf2dcb-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-5j9vm\" (UID: \"7ef4b609-a02d-4050-a5cc-768308bf2dcb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm" Dec 02 01:51:33 crc kubenswrapper[4884]: I1202 01:51:33.341126 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-memberlist\") pod \"speaker-bxtbh\" (UID: \"f3fc51b3-99fe-4f7f-be90-68acd111b7ca\") " pod="metallb-system/speaker-bxtbh" Dec 02 01:51:33 crc kubenswrapper[4884]: I1202 01:51:33.341188 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-metrics-certs\") pod \"speaker-bxtbh\" (UID: \"f3fc51b3-99fe-4f7f-be90-68acd111b7ca\") " pod="metallb-system/speaker-bxtbh" Dec 02 01:51:33 crc kubenswrapper[4884]: I1202 01:51:33.341219 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e365201e-b871-4206-8dcd-9b87a9452f83-metrics-certs\") pod \"controller-f8648f98b-8wv2n\" (UID: \"e365201e-b871-4206-8dcd-9b87a9452f83\") " pod="metallb-system/controller-f8648f98b-8wv2n" Dec 02 01:51:33 crc kubenswrapper[4884]: E1202 01:51:33.341313 4884 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 02 01:51:33 crc kubenswrapper[4884]: E1202 01:51:33.341407 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-memberlist podName:f3fc51b3-99fe-4f7f-be90-68acd111b7ca nodeName:}" failed. No retries permitted until 2025-12-02 01:51:34.341388384 +0000 UTC m=+911.017225268 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-memberlist") pod "speaker-bxtbh" (UID: "f3fc51b3-99fe-4f7f-be90-68acd111b7ca") : secret "metallb-memberlist" not found Dec 02 01:51:33 crc kubenswrapper[4884]: I1202 01:51:33.343972 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-metrics-certs\") pod \"speaker-bxtbh\" (UID: \"f3fc51b3-99fe-4f7f-be90-68acd111b7ca\") " pod="metallb-system/speaker-bxtbh" Dec 02 01:51:33 crc kubenswrapper[4884]: I1202 01:51:33.344246 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e365201e-b871-4206-8dcd-9b87a9452f83-metrics-certs\") pod \"controller-f8648f98b-8wv2n\" (UID: \"e365201e-b871-4206-8dcd-9b87a9452f83\") " pod="metallb-system/controller-f8648f98b-8wv2n" Dec 02 01:51:33 crc kubenswrapper[4884]: I1202 01:51:33.367167 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-8wv2n" Dec 02 01:51:33 crc kubenswrapper[4884]: I1202 01:51:33.512056 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm" Dec 02 01:51:33 crc kubenswrapper[4884]: I1202 01:51:33.596781 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-8wv2n"] Dec 02 01:51:33 crc kubenswrapper[4884]: W1202 01:51:33.604063 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode365201e_b871_4206_8dcd_9b87a9452f83.slice/crio-0f48040c6bfd388e7fb6138c955aa98e44ecf2fb912960639b654343b5cfe041 WatchSource:0}: Error finding container 0f48040c6bfd388e7fb6138c955aa98e44ecf2fb912960639b654343b5cfe041: Status 404 returned error can't find the container with id 0f48040c6bfd388e7fb6138c955aa98e44ecf2fb912960639b654343b5cfe041 Dec 02 01:51:33 crc kubenswrapper[4884]: I1202 01:51:33.914965 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm"] Dec 02 01:51:33 crc kubenswrapper[4884]: W1202 01:51:33.924029 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ef4b609_a02d_4050_a5cc_768308bf2dcb.slice/crio-6d1d6efa439e6b10580c63a00710c6a01063455d9409d58282a3804d4ba1b2e6 WatchSource:0}: Error finding container 6d1d6efa439e6b10580c63a00710c6a01063455d9409d58282a3804d4ba1b2e6: Status 404 returned error can't find the container with id 6d1d6efa439e6b10580c63a00710c6a01063455d9409d58282a3804d4ba1b2e6 Dec 02 01:51:34 crc kubenswrapper[4884]: I1202 01:51:34.059082 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7xs4" event={"ID":"d0e2e152-8ae0-495d-835c-5dca5ef66b6c","Type":"ContainerStarted","Data":"fbc485c5e568f028f742f27847e7c6ff619bb8ea4c80ba4db9f30fe50a79ed46"} Dec 02 01:51:34 crc kubenswrapper[4884]: I1202 01:51:34.060026 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm" event={"ID":"7ef4b609-a02d-4050-a5cc-768308bf2dcb","Type":"ContainerStarted","Data":"6d1d6efa439e6b10580c63a00710c6a01063455d9409d58282a3804d4ba1b2e6"} Dec 02 01:51:34 crc kubenswrapper[4884]: I1202 01:51:34.061360 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-8wv2n" event={"ID":"e365201e-b871-4206-8dcd-9b87a9452f83","Type":"ContainerStarted","Data":"d251b85cfb891410b856b0eb8396859c45b40ba15550ccc2c016c631d9caa3d9"} Dec 02 01:51:34 crc kubenswrapper[4884]: I1202 01:51:34.061384 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-8wv2n" event={"ID":"e365201e-b871-4206-8dcd-9b87a9452f83","Type":"ContainerStarted","Data":"b3b93fe63cef152a48069846ee356362ac6f1d1ed647af84167182d1cedb4ff2"} Dec 02 01:51:34 crc kubenswrapper[4884]: I1202 01:51:34.061394 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-8wv2n" event={"ID":"e365201e-b871-4206-8dcd-9b87a9452f83","Type":"ContainerStarted","Data":"0f48040c6bfd388e7fb6138c955aa98e44ecf2fb912960639b654343b5cfe041"} Dec 02 01:51:34 crc kubenswrapper[4884]: I1202 01:51:34.061545 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-8wv2n" Dec 02 01:51:34 crc kubenswrapper[4884]: I1202 01:51:34.082381 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-8wv2n" podStartSLOduration=2.082364561 podStartE2EDuration="2.082364561s" podCreationTimestamp="2025-12-02 01:51:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:51:34.076892105 +0000 UTC m=+910.752729009" watchObservedRunningTime="2025-12-02 01:51:34.082364561 +0000 UTC m=+910.758201455" Dec 02 01:51:34 crc kubenswrapper[4884]: I1202 01:51:34.364487 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-memberlist\") pod \"speaker-bxtbh\" (UID: \"f3fc51b3-99fe-4f7f-be90-68acd111b7ca\") " pod="metallb-system/speaker-bxtbh" Dec 02 01:51:34 crc kubenswrapper[4884]: I1202 01:51:34.386675 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/f3fc51b3-99fe-4f7f-be90-68acd111b7ca-memberlist\") pod \"speaker-bxtbh\" (UID: \"f3fc51b3-99fe-4f7f-be90-68acd111b7ca\") " pod="metallb-system/speaker-bxtbh" Dec 02 01:51:34 crc kubenswrapper[4884]: I1202 01:51:34.525433 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-bxtbh" Dec 02 01:51:35 crc kubenswrapper[4884]: I1202 01:51:35.073034 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bxtbh" event={"ID":"f3fc51b3-99fe-4f7f-be90-68acd111b7ca","Type":"ContainerStarted","Data":"44a83ce3dc132d1ff0c926d2ef016ef0cce22ba76bdadbb1ee58181753f2a44c"} Dec 02 01:51:35 crc kubenswrapper[4884]: I1202 01:51:35.073638 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bxtbh" event={"ID":"f3fc51b3-99fe-4f7f-be90-68acd111b7ca","Type":"ContainerStarted","Data":"d333bd16b3f5146740f0bf41384358543f6091c19bbe3f9e763a45060c93011f"} Dec 02 01:51:36 crc kubenswrapper[4884]: I1202 01:51:36.101029 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-bxtbh" event={"ID":"f3fc51b3-99fe-4f7f-be90-68acd111b7ca","Type":"ContainerStarted","Data":"b121477c356ccfbf7ac9776871c2e01c07181646cfcc577bdc3efe91df4162bd"} Dec 02 01:51:36 crc kubenswrapper[4884]: I1202 01:51:36.101825 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-bxtbh" Dec 02 01:51:36 crc kubenswrapper[4884]: I1202 01:51:36.130248 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-bxtbh" podStartSLOduration=4.130222605 podStartE2EDuration="4.130222605s" podCreationTimestamp="2025-12-02 01:51:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:51:36.122170525 +0000 UTC m=+912.798007409" watchObservedRunningTime="2025-12-02 01:51:36.130222605 +0000 UTC m=+912.806059489" Dec 02 01:51:41 crc kubenswrapper[4884]: I1202 01:51:41.142581 4884 generic.go:334] "Generic (PLEG): container finished" podID="d0e2e152-8ae0-495d-835c-5dca5ef66b6c" containerID="2d64e8673f3d12ae1a16f3e3909c7b0006b59b48457f9743bf2395cbe699b0f8" exitCode=0 Dec 02 01:51:41 crc kubenswrapper[4884]: I1202 01:51:41.142821 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7xs4" event={"ID":"d0e2e152-8ae0-495d-835c-5dca5ef66b6c","Type":"ContainerDied","Data":"2d64e8673f3d12ae1a16f3e3909c7b0006b59b48457f9743bf2395cbe699b0f8"} Dec 02 01:51:41 crc kubenswrapper[4884]: I1202 01:51:41.145343 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm" event={"ID":"7ef4b609-a02d-4050-a5cc-768308bf2dcb","Type":"ContainerStarted","Data":"8a76ed2642f39622b3551a7d8d261fc92e1288a88967b29de7fc09278f1b4252"} Dec 02 01:51:41 crc kubenswrapper[4884]: I1202 01:51:41.145536 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm" Dec 02 01:51:41 crc kubenswrapper[4884]: I1202 01:51:41.223126 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm" podStartSLOduration=2.6544837550000002 podStartE2EDuration="9.223094666s" podCreationTimestamp="2025-12-02 01:51:32 +0000 UTC" firstStartedPulling="2025-12-02 01:51:33.926437049 +0000 UTC m=+910.602273933" lastFinishedPulling="2025-12-02 01:51:40.49504796 +0000 UTC m=+917.170884844" observedRunningTime="2025-12-02 01:51:41.212205915 +0000 UTC m=+917.888042839" watchObservedRunningTime="2025-12-02 01:51:41.223094666 +0000 UTC m=+917.898931580" Dec 02 01:51:42 crc kubenswrapper[4884]: I1202 01:51:42.154262 4884 generic.go:334] "Generic (PLEG): container finished" podID="d0e2e152-8ae0-495d-835c-5dca5ef66b6c" containerID="998b818ddadcf00d8c57944d500b14ab7b63c960f31081a58dc2934025784327" exitCode=0 Dec 02 01:51:42 crc kubenswrapper[4884]: I1202 01:51:42.154374 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7xs4" event={"ID":"d0e2e152-8ae0-495d-835c-5dca5ef66b6c","Type":"ContainerDied","Data":"998b818ddadcf00d8c57944d500b14ab7b63c960f31081a58dc2934025784327"} Dec 02 01:51:43 crc kubenswrapper[4884]: I1202 01:51:43.379076 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-8wv2n" Dec 02 01:51:44 crc kubenswrapper[4884]: I1202 01:51:44.325400 4884 generic.go:334] "Generic (PLEG): container finished" podID="d0e2e152-8ae0-495d-835c-5dca5ef66b6c" containerID="0d88f7de38f7132cef05964a17eb95973eb6f0ecd9a3d8209bceba40ccf3ffa7" exitCode=0 Dec 02 01:51:44 crc kubenswrapper[4884]: I1202 01:51:44.325462 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7xs4" event={"ID":"d0e2e152-8ae0-495d-835c-5dca5ef66b6c","Type":"ContainerDied","Data":"0d88f7de38f7132cef05964a17eb95973eb6f0ecd9a3d8209bceba40ccf3ffa7"} Dec 02 01:51:44 crc kubenswrapper[4884]: I1202 01:51:44.530836 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-bxtbh" Dec 02 01:51:45 crc kubenswrapper[4884]: I1202 01:51:45.340286 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7xs4" event={"ID":"d0e2e152-8ae0-495d-835c-5dca5ef66b6c","Type":"ContainerStarted","Data":"63627a71ca27d30f0f79e6bebb6acc3ae5be6ae47a931460fada6eed8a9ee3ce"} Dec 02 01:51:45 crc kubenswrapper[4884]: I1202 01:51:45.340638 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7xs4" event={"ID":"d0e2e152-8ae0-495d-835c-5dca5ef66b6c","Type":"ContainerStarted","Data":"f56c99ff7478221f236b6359ca8c8872291f7c2c8786c790bba2bea097ced9a0"} Dec 02 01:51:45 crc kubenswrapper[4884]: I1202 01:51:45.340651 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7xs4" event={"ID":"d0e2e152-8ae0-495d-835c-5dca5ef66b6c","Type":"ContainerStarted","Data":"fc080b464440a9c0565d515dc4b6887e8684ef04c4abd1afeccaf40299706ae6"} Dec 02 01:51:45 crc kubenswrapper[4884]: I1202 01:51:45.340665 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7xs4" event={"ID":"d0e2e152-8ae0-495d-835c-5dca5ef66b6c","Type":"ContainerStarted","Data":"a6bc9a86e381a2c348f9ebbb594898b39b3bbf85fb4ac82f418f2b8695489f4c"} Dec 02 01:51:45 crc kubenswrapper[4884]: I1202 01:51:45.340699 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7xs4" event={"ID":"d0e2e152-8ae0-495d-835c-5dca5ef66b6c","Type":"ContainerStarted","Data":"e5d943bdafe3909125705f5ae76c7a7d2f2398a69c920a1bed8278c2de1c6af4"} Dec 02 01:51:46 crc kubenswrapper[4884]: I1202 01:51:46.354182 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-j7xs4" event={"ID":"d0e2e152-8ae0-495d-835c-5dca5ef66b6c","Type":"ContainerStarted","Data":"cb64ba36b79a5990829165b61aa4ce3f629f4d0a2f64e803aa70b78c321bf957"} Dec 02 01:51:46 crc kubenswrapper[4884]: I1202 01:51:46.354581 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:46 crc kubenswrapper[4884]: I1202 01:51:46.381282 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-j7xs4" podStartSLOduration=6.993291507 podStartE2EDuration="14.381261349s" podCreationTimestamp="2025-12-02 01:51:32 +0000 UTC" firstStartedPulling="2025-12-02 01:51:33.074378296 +0000 UTC m=+909.750215180" lastFinishedPulling="2025-12-02 01:51:40.462348138 +0000 UTC m=+917.138185022" observedRunningTime="2025-12-02 01:51:46.378309626 +0000 UTC m=+923.054146550" watchObservedRunningTime="2025-12-02 01:51:46.381261349 +0000 UTC m=+923.057098233" Dec 02 01:51:47 crc kubenswrapper[4884]: I1202 01:51:47.638161 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-k6qf9"] Dec 02 01:51:47 crc kubenswrapper[4884]: I1202 01:51:47.639674 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k6qf9" Dec 02 01:51:47 crc kubenswrapper[4884]: I1202 01:51:47.642679 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-7n9jq" Dec 02 01:51:47 crc kubenswrapper[4884]: I1202 01:51:47.643325 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 02 01:51:47 crc kubenswrapper[4884]: I1202 01:51:47.644332 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 02 01:51:47 crc kubenswrapper[4884]: I1202 01:51:47.664857 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-k6qf9"] Dec 02 01:51:47 crc kubenswrapper[4884]: I1202 01:51:47.766373 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6zvf\" (UniqueName: \"kubernetes.io/projected/6db7318b-7a76-4b95-b2bb-14a2dbd312b6-kube-api-access-k6zvf\") pod \"openstack-operator-index-k6qf9\" (UID: \"6db7318b-7a76-4b95-b2bb-14a2dbd312b6\") " pod="openstack-operators/openstack-operator-index-k6qf9" Dec 02 01:51:47 crc kubenswrapper[4884]: I1202 01:51:47.867767 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6zvf\" (UniqueName: \"kubernetes.io/projected/6db7318b-7a76-4b95-b2bb-14a2dbd312b6-kube-api-access-k6zvf\") pod \"openstack-operator-index-k6qf9\" (UID: \"6db7318b-7a76-4b95-b2bb-14a2dbd312b6\") " pod="openstack-operators/openstack-operator-index-k6qf9" Dec 02 01:51:47 crc kubenswrapper[4884]: I1202 01:51:47.894506 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6zvf\" (UniqueName: \"kubernetes.io/projected/6db7318b-7a76-4b95-b2bb-14a2dbd312b6-kube-api-access-k6zvf\") pod \"openstack-operator-index-k6qf9\" (UID: \"6db7318b-7a76-4b95-b2bb-14a2dbd312b6\") " pod="openstack-operators/openstack-operator-index-k6qf9" Dec 02 01:51:47 crc kubenswrapper[4884]: I1202 01:51:47.919520 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:47 crc kubenswrapper[4884]: I1202 01:51:47.953613 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:51:47 crc kubenswrapper[4884]: I1202 01:51:47.972276 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k6qf9" Dec 02 01:51:48 crc kubenswrapper[4884]: I1202 01:51:48.461967 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-k6qf9"] Dec 02 01:51:49 crc kubenswrapper[4884]: I1202 01:51:49.378365 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k6qf9" event={"ID":"6db7318b-7a76-4b95-b2bb-14a2dbd312b6","Type":"ContainerStarted","Data":"ace2b3802b83e7ffa1d1dd2e4143bda0bec978f7c060e06f06331321c0ef1e85"} Dec 02 01:51:50 crc kubenswrapper[4884]: I1202 01:51:50.993296 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-k6qf9"] Dec 02 01:51:51 crc kubenswrapper[4884]: I1202 01:51:51.596972 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-rh4qh"] Dec 02 01:51:51 crc kubenswrapper[4884]: I1202 01:51:51.598119 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rh4qh" Dec 02 01:51:51 crc kubenswrapper[4884]: I1202 01:51:51.627821 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rh4qh"] Dec 02 01:51:51 crc kubenswrapper[4884]: I1202 01:51:51.721782 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nbkt\" (UniqueName: \"kubernetes.io/projected/6eb2895e-8fa8-4e8d-b171-0c47bbccf248-kube-api-access-8nbkt\") pod \"openstack-operator-index-rh4qh\" (UID: \"6eb2895e-8fa8-4e8d-b171-0c47bbccf248\") " pod="openstack-operators/openstack-operator-index-rh4qh" Dec 02 01:51:51 crc kubenswrapper[4884]: I1202 01:51:51.823078 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nbkt\" (UniqueName: \"kubernetes.io/projected/6eb2895e-8fa8-4e8d-b171-0c47bbccf248-kube-api-access-8nbkt\") pod \"openstack-operator-index-rh4qh\" (UID: \"6eb2895e-8fa8-4e8d-b171-0c47bbccf248\") " pod="openstack-operators/openstack-operator-index-rh4qh" Dec 02 01:51:51 crc kubenswrapper[4884]: I1202 01:51:51.850862 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nbkt\" (UniqueName: \"kubernetes.io/projected/6eb2895e-8fa8-4e8d-b171-0c47bbccf248-kube-api-access-8nbkt\") pod \"openstack-operator-index-rh4qh\" (UID: \"6eb2895e-8fa8-4e8d-b171-0c47bbccf248\") " pod="openstack-operators/openstack-operator-index-rh4qh" Dec 02 01:51:51 crc kubenswrapper[4884]: I1202 01:51:51.956162 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rh4qh" Dec 02 01:51:52 crc kubenswrapper[4884]: I1202 01:51:52.230238 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rh4qh"] Dec 02 01:51:52 crc kubenswrapper[4884]: W1202 01:51:52.237290 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6eb2895e_8fa8_4e8d_b171_0c47bbccf248.slice/crio-4c88434b59f9620cceb72b64a614f0eabeaab519cb5bb1360f0df042dee372d7 WatchSource:0}: Error finding container 4c88434b59f9620cceb72b64a614f0eabeaab519cb5bb1360f0df042dee372d7: Status 404 returned error can't find the container with id 4c88434b59f9620cceb72b64a614f0eabeaab519cb5bb1360f0df042dee372d7 Dec 02 01:51:52 crc kubenswrapper[4884]: I1202 01:51:52.404478 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k6qf9" event={"ID":"6db7318b-7a76-4b95-b2bb-14a2dbd312b6","Type":"ContainerStarted","Data":"0d81128175c1ff1cbf2ac95d6d5e876be3262e539b8d145df1484dd0b957df83"} Dec 02 01:51:52 crc kubenswrapper[4884]: I1202 01:51:52.404577 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-k6qf9" podUID="6db7318b-7a76-4b95-b2bb-14a2dbd312b6" containerName="registry-server" containerID="cri-o://0d81128175c1ff1cbf2ac95d6d5e876be3262e539b8d145df1484dd0b957df83" gracePeriod=2 Dec 02 01:51:52 crc kubenswrapper[4884]: I1202 01:51:52.406101 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rh4qh" event={"ID":"6eb2895e-8fa8-4e8d-b171-0c47bbccf248","Type":"ContainerStarted","Data":"4c88434b59f9620cceb72b64a614f0eabeaab519cb5bb1360f0df042dee372d7"} Dec 02 01:51:52 crc kubenswrapper[4884]: I1202 01:51:52.428348 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-k6qf9" podStartSLOduration=2.393912065 podStartE2EDuration="5.42832116s" podCreationTimestamp="2025-12-02 01:51:47 +0000 UTC" firstStartedPulling="2025-12-02 01:51:48.472400826 +0000 UTC m=+925.148237750" lastFinishedPulling="2025-12-02 01:51:51.506809951 +0000 UTC m=+928.182646845" observedRunningTime="2025-12-02 01:51:52.42549014 +0000 UTC m=+929.101327064" watchObservedRunningTime="2025-12-02 01:51:52.42832116 +0000 UTC m=+929.104158084" Dec 02 01:51:52 crc kubenswrapper[4884]: I1202 01:51:52.850173 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k6qf9" Dec 02 01:51:52 crc kubenswrapper[4884]: I1202 01:51:52.937637 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6zvf\" (UniqueName: \"kubernetes.io/projected/6db7318b-7a76-4b95-b2bb-14a2dbd312b6-kube-api-access-k6zvf\") pod \"6db7318b-7a76-4b95-b2bb-14a2dbd312b6\" (UID: \"6db7318b-7a76-4b95-b2bb-14a2dbd312b6\") " Dec 02 01:51:52 crc kubenswrapper[4884]: I1202 01:51:52.946729 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6db7318b-7a76-4b95-b2bb-14a2dbd312b6-kube-api-access-k6zvf" (OuterVolumeSpecName: "kube-api-access-k6zvf") pod "6db7318b-7a76-4b95-b2bb-14a2dbd312b6" (UID: "6db7318b-7a76-4b95-b2bb-14a2dbd312b6"). InnerVolumeSpecName "kube-api-access-k6zvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:51:53 crc kubenswrapper[4884]: I1202 01:51:53.039438 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6zvf\" (UniqueName: \"kubernetes.io/projected/6db7318b-7a76-4b95-b2bb-14a2dbd312b6-kube-api-access-k6zvf\") on node \"crc\" DevicePath \"\"" Dec 02 01:51:53 crc kubenswrapper[4884]: I1202 01:51:53.417714 4884 generic.go:334] "Generic (PLEG): container finished" podID="6db7318b-7a76-4b95-b2bb-14a2dbd312b6" containerID="0d81128175c1ff1cbf2ac95d6d5e876be3262e539b8d145df1484dd0b957df83" exitCode=0 Dec 02 01:51:53 crc kubenswrapper[4884]: I1202 01:51:53.417791 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k6qf9" event={"ID":"6db7318b-7a76-4b95-b2bb-14a2dbd312b6","Type":"ContainerDied","Data":"0d81128175c1ff1cbf2ac95d6d5e876be3262e539b8d145df1484dd0b957df83"} Dec 02 01:51:53 crc kubenswrapper[4884]: I1202 01:51:53.417836 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k6qf9" event={"ID":"6db7318b-7a76-4b95-b2bb-14a2dbd312b6","Type":"ContainerDied","Data":"ace2b3802b83e7ffa1d1dd2e4143bda0bec978f7c060e06f06331321c0ef1e85"} Dec 02 01:51:53 crc kubenswrapper[4884]: I1202 01:51:53.417855 4884 scope.go:117] "RemoveContainer" containerID="0d81128175c1ff1cbf2ac95d6d5e876be3262e539b8d145df1484dd0b957df83" Dec 02 01:51:53 crc kubenswrapper[4884]: I1202 01:51:53.417870 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k6qf9" Dec 02 01:51:53 crc kubenswrapper[4884]: I1202 01:51:53.420017 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rh4qh" event={"ID":"6eb2895e-8fa8-4e8d-b171-0c47bbccf248","Type":"ContainerStarted","Data":"cd4ae6bcabef24fdfe6f7825e4136f2423214079bad4450a01ed4d07c02a970f"} Dec 02 01:51:53 crc kubenswrapper[4884]: I1202 01:51:53.445939 4884 scope.go:117] "RemoveContainer" containerID="0d81128175c1ff1cbf2ac95d6d5e876be3262e539b8d145df1484dd0b957df83" Dec 02 01:51:53 crc kubenswrapper[4884]: E1202 01:51:53.446551 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d81128175c1ff1cbf2ac95d6d5e876be3262e539b8d145df1484dd0b957df83\": container with ID starting with 0d81128175c1ff1cbf2ac95d6d5e876be3262e539b8d145df1484dd0b957df83 not found: ID does not exist" containerID="0d81128175c1ff1cbf2ac95d6d5e876be3262e539b8d145df1484dd0b957df83" Dec 02 01:51:53 crc kubenswrapper[4884]: I1202 01:51:53.446582 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d81128175c1ff1cbf2ac95d6d5e876be3262e539b8d145df1484dd0b957df83"} err="failed to get container status \"0d81128175c1ff1cbf2ac95d6d5e876be3262e539b8d145df1484dd0b957df83\": rpc error: code = NotFound desc = could not find container \"0d81128175c1ff1cbf2ac95d6d5e876be3262e539b8d145df1484dd0b957df83\": container with ID starting with 0d81128175c1ff1cbf2ac95d6d5e876be3262e539b8d145df1484dd0b957df83 not found: ID does not exist" Dec 02 01:51:53 crc kubenswrapper[4884]: I1202 01:51:53.451922 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-rh4qh" podStartSLOduration=2.391215476 podStartE2EDuration="2.451869802s" podCreationTimestamp="2025-12-02 01:51:51 +0000 UTC" firstStartedPulling="2025-12-02 01:51:52.241998424 +0000 UTC m=+928.917835318" lastFinishedPulling="2025-12-02 01:51:52.30265276 +0000 UTC m=+928.978489644" observedRunningTime="2025-12-02 01:51:53.44571258 +0000 UTC m=+930.121549524" watchObservedRunningTime="2025-12-02 01:51:53.451869802 +0000 UTC m=+930.127706746" Dec 02 01:51:53 crc kubenswrapper[4884]: I1202 01:51:53.475146 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-k6qf9"] Dec 02 01:51:53 crc kubenswrapper[4884]: I1202 01:51:53.484383 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-k6qf9"] Dec 02 01:51:53 crc kubenswrapper[4884]: I1202 01:51:53.522242 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5j9vm" Dec 02 01:51:53 crc kubenswrapper[4884]: I1202 01:51:53.624354 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6db7318b-7a76-4b95-b2bb-14a2dbd312b6" path="/var/lib/kubelet/pods/6db7318b-7a76-4b95-b2bb-14a2dbd312b6/volumes" Dec 02 01:52:01 crc kubenswrapper[4884]: I1202 01:52:01.957222 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-rh4qh" Dec 02 01:52:01 crc kubenswrapper[4884]: I1202 01:52:01.958058 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-rh4qh" Dec 02 01:52:02 crc kubenswrapper[4884]: I1202 01:52:02.000148 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-rh4qh" Dec 02 01:52:02 crc kubenswrapper[4884]: I1202 01:52:02.561998 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-rh4qh" Dec 02 01:52:02 crc kubenswrapper[4884]: I1202 01:52:02.923204 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-j7xs4" Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.190905 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs"] Dec 02 01:52:08 crc kubenswrapper[4884]: E1202 01:52:08.192013 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6db7318b-7a76-4b95-b2bb-14a2dbd312b6" containerName="registry-server" Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.192031 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="6db7318b-7a76-4b95-b2bb-14a2dbd312b6" containerName="registry-server" Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.194986 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="6db7318b-7a76-4b95-b2bb-14a2dbd312b6" containerName="registry-server" Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.199133 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.210128 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-rhxrw" Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.220872 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs"] Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.381719 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-util\") pod \"290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs\" (UID: \"0dc571b2-e65a-41d5-b47c-8acbb09cbe28\") " pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.381859 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-bundle\") pod \"290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs\" (UID: \"0dc571b2-e65a-41d5-b47c-8acbb09cbe28\") " pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.382043 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk7n7\" (UniqueName: \"kubernetes.io/projected/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-kube-api-access-tk7n7\") pod \"290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs\" (UID: \"0dc571b2-e65a-41d5-b47c-8acbb09cbe28\") " pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.484097 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-bundle\") pod \"290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs\" (UID: \"0dc571b2-e65a-41d5-b47c-8acbb09cbe28\") " pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.484229 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk7n7\" (UniqueName: \"kubernetes.io/projected/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-kube-api-access-tk7n7\") pod \"290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs\" (UID: \"0dc571b2-e65a-41d5-b47c-8acbb09cbe28\") " pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.484314 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-util\") pod \"290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs\" (UID: \"0dc571b2-e65a-41d5-b47c-8acbb09cbe28\") " pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.484630 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-bundle\") pod \"290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs\" (UID: \"0dc571b2-e65a-41d5-b47c-8acbb09cbe28\") " pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.484964 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-util\") pod \"290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs\" (UID: \"0dc571b2-e65a-41d5-b47c-8acbb09cbe28\") " pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.521060 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk7n7\" (UniqueName: \"kubernetes.io/projected/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-kube-api-access-tk7n7\") pod \"290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs\" (UID: \"0dc571b2-e65a-41d5-b47c-8acbb09cbe28\") " pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" Dec 02 01:52:08 crc kubenswrapper[4884]: I1202 01:52:08.529028 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" Dec 02 01:52:09 crc kubenswrapper[4884]: I1202 01:52:09.096311 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs"] Dec 02 01:52:09 crc kubenswrapper[4884]: I1202 01:52:09.556493 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" event={"ID":"0dc571b2-e65a-41d5-b47c-8acbb09cbe28","Type":"ContainerStarted","Data":"9bb2ac8769b6098e0fa5b59418585af7519abd94b4d08562c2c13b1f1eeb1935"} Dec 02 01:52:10 crc kubenswrapper[4884]: I1202 01:52:10.563001 4884 generic.go:334] "Generic (PLEG): container finished" podID="0dc571b2-e65a-41d5-b47c-8acbb09cbe28" containerID="dd8e42be665bb50fb5916678a7a85dc40684e3378d844d5261c01b4da3743c4d" exitCode=0 Dec 02 01:52:10 crc kubenswrapper[4884]: I1202 01:52:10.563068 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" event={"ID":"0dc571b2-e65a-41d5-b47c-8acbb09cbe28","Type":"ContainerDied","Data":"dd8e42be665bb50fb5916678a7a85dc40684e3378d844d5261c01b4da3743c4d"} Dec 02 01:52:11 crc kubenswrapper[4884]: I1202 01:52:11.574234 4884 generic.go:334] "Generic (PLEG): container finished" podID="0dc571b2-e65a-41d5-b47c-8acbb09cbe28" containerID="433e2e86e2ba4d677c8fe99d9d8a90d1acd4b063e0114ed80686c1581c061712" exitCode=0 Dec 02 01:52:11 crc kubenswrapper[4884]: I1202 01:52:11.574349 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" event={"ID":"0dc571b2-e65a-41d5-b47c-8acbb09cbe28","Type":"ContainerDied","Data":"433e2e86e2ba4d677c8fe99d9d8a90d1acd4b063e0114ed80686c1581c061712"} Dec 02 01:52:12 crc kubenswrapper[4884]: I1202 01:52:12.588121 4884 generic.go:334] "Generic (PLEG): container finished" podID="0dc571b2-e65a-41d5-b47c-8acbb09cbe28" containerID="b548ed11574c39d192b838b704c484bd74c1e63cff27b88abcdf2e9fa52d4cd2" exitCode=0 Dec 02 01:52:12 crc kubenswrapper[4884]: I1202 01:52:12.588170 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" event={"ID":"0dc571b2-e65a-41d5-b47c-8acbb09cbe28","Type":"ContainerDied","Data":"b548ed11574c39d192b838b704c484bd74c1e63cff27b88abcdf2e9fa52d4cd2"} Dec 02 01:52:14 crc kubenswrapper[4884]: I1202 01:52:14.024479 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" Dec 02 01:52:14 crc kubenswrapper[4884]: I1202 01:52:14.076673 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-util\") pod \"0dc571b2-e65a-41d5-b47c-8acbb09cbe28\" (UID: \"0dc571b2-e65a-41d5-b47c-8acbb09cbe28\") " Dec 02 01:52:14 crc kubenswrapper[4884]: I1202 01:52:14.076779 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-bundle\") pod \"0dc571b2-e65a-41d5-b47c-8acbb09cbe28\" (UID: \"0dc571b2-e65a-41d5-b47c-8acbb09cbe28\") " Dec 02 01:52:14 crc kubenswrapper[4884]: I1202 01:52:14.076820 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk7n7\" (UniqueName: \"kubernetes.io/projected/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-kube-api-access-tk7n7\") pod \"0dc571b2-e65a-41d5-b47c-8acbb09cbe28\" (UID: \"0dc571b2-e65a-41d5-b47c-8acbb09cbe28\") " Dec 02 01:52:14 crc kubenswrapper[4884]: I1202 01:52:14.077456 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-bundle" (OuterVolumeSpecName: "bundle") pod "0dc571b2-e65a-41d5-b47c-8acbb09cbe28" (UID: "0dc571b2-e65a-41d5-b47c-8acbb09cbe28"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:52:14 crc kubenswrapper[4884]: I1202 01:52:14.082546 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-kube-api-access-tk7n7" (OuterVolumeSpecName: "kube-api-access-tk7n7") pod "0dc571b2-e65a-41d5-b47c-8acbb09cbe28" (UID: "0dc571b2-e65a-41d5-b47c-8acbb09cbe28"). InnerVolumeSpecName "kube-api-access-tk7n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:52:14 crc kubenswrapper[4884]: I1202 01:52:14.106541 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-util" (OuterVolumeSpecName: "util") pod "0dc571b2-e65a-41d5-b47c-8acbb09cbe28" (UID: "0dc571b2-e65a-41d5-b47c-8acbb09cbe28"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:52:14 crc kubenswrapper[4884]: I1202 01:52:14.178847 4884 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-util\") on node \"crc\" DevicePath \"\"" Dec 02 01:52:14 crc kubenswrapper[4884]: I1202 01:52:14.178897 4884 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:52:14 crc kubenswrapper[4884]: I1202 01:52:14.178917 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk7n7\" (UniqueName: \"kubernetes.io/projected/0dc571b2-e65a-41d5-b47c-8acbb09cbe28-kube-api-access-tk7n7\") on node \"crc\" DevicePath \"\"" Dec 02 01:52:14 crc kubenswrapper[4884]: I1202 01:52:14.606465 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" event={"ID":"0dc571b2-e65a-41d5-b47c-8acbb09cbe28","Type":"ContainerDied","Data":"9bb2ac8769b6098e0fa5b59418585af7519abd94b4d08562c2c13b1f1eeb1935"} Dec 02 01:52:14 crc kubenswrapper[4884]: I1202 01:52:14.606505 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bb2ac8769b6098e0fa5b59418585af7519abd94b4d08562c2c13b1f1eeb1935" Dec 02 01:52:14 crc kubenswrapper[4884]: I1202 01:52:14.606556 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs" Dec 02 01:52:22 crc kubenswrapper[4884]: I1202 01:52:22.483093 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-557b87d65-9tj59"] Dec 02 01:52:22 crc kubenswrapper[4884]: E1202 01:52:22.491805 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dc571b2-e65a-41d5-b47c-8acbb09cbe28" containerName="pull" Dec 02 01:52:22 crc kubenswrapper[4884]: I1202 01:52:22.491835 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dc571b2-e65a-41d5-b47c-8acbb09cbe28" containerName="pull" Dec 02 01:52:22 crc kubenswrapper[4884]: E1202 01:52:22.491853 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dc571b2-e65a-41d5-b47c-8acbb09cbe28" containerName="extract" Dec 02 01:52:22 crc kubenswrapper[4884]: I1202 01:52:22.491862 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dc571b2-e65a-41d5-b47c-8acbb09cbe28" containerName="extract" Dec 02 01:52:22 crc kubenswrapper[4884]: E1202 01:52:22.491876 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dc571b2-e65a-41d5-b47c-8acbb09cbe28" containerName="util" Dec 02 01:52:22 crc kubenswrapper[4884]: I1202 01:52:22.491884 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dc571b2-e65a-41d5-b47c-8acbb09cbe28" containerName="util" Dec 02 01:52:22 crc kubenswrapper[4884]: I1202 01:52:22.492027 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dc571b2-e65a-41d5-b47c-8acbb09cbe28" containerName="extract" Dec 02 01:52:22 crc kubenswrapper[4884]: I1202 01:52:22.492529 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-557b87d65-9tj59" Dec 02 01:52:22 crc kubenswrapper[4884]: I1202 01:52:22.499050 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-klzr6" Dec 02 01:52:22 crc kubenswrapper[4884]: I1202 01:52:22.499991 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfslq\" (UniqueName: \"kubernetes.io/projected/4f2e27ae-6dcb-4d32-b65b-f7d96a026d25-kube-api-access-gfslq\") pod \"openstack-operator-controller-operator-557b87d65-9tj59\" (UID: \"4f2e27ae-6dcb-4d32-b65b-f7d96a026d25\") " pod="openstack-operators/openstack-operator-controller-operator-557b87d65-9tj59" Dec 02 01:52:22 crc kubenswrapper[4884]: I1202 01:52:22.534904 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-557b87d65-9tj59"] Dec 02 01:52:22 crc kubenswrapper[4884]: I1202 01:52:22.601500 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfslq\" (UniqueName: \"kubernetes.io/projected/4f2e27ae-6dcb-4d32-b65b-f7d96a026d25-kube-api-access-gfslq\") pod \"openstack-operator-controller-operator-557b87d65-9tj59\" (UID: \"4f2e27ae-6dcb-4d32-b65b-f7d96a026d25\") " pod="openstack-operators/openstack-operator-controller-operator-557b87d65-9tj59" Dec 02 01:52:22 crc kubenswrapper[4884]: I1202 01:52:22.627496 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfslq\" (UniqueName: \"kubernetes.io/projected/4f2e27ae-6dcb-4d32-b65b-f7d96a026d25-kube-api-access-gfslq\") pod \"openstack-operator-controller-operator-557b87d65-9tj59\" (UID: \"4f2e27ae-6dcb-4d32-b65b-f7d96a026d25\") " pod="openstack-operators/openstack-operator-controller-operator-557b87d65-9tj59" Dec 02 01:52:22 crc kubenswrapper[4884]: I1202 01:52:22.810558 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-557b87d65-9tj59" Dec 02 01:52:23 crc kubenswrapper[4884]: I1202 01:52:23.328358 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-557b87d65-9tj59"] Dec 02 01:52:23 crc kubenswrapper[4884]: I1202 01:52:23.672761 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-557b87d65-9tj59" event={"ID":"4f2e27ae-6dcb-4d32-b65b-f7d96a026d25","Type":"ContainerStarted","Data":"34749fc6f35c40e9556d61769e8cd1c2ab9fb118830391bfd2c13f27fdf23173"} Dec 02 01:52:27 crc kubenswrapper[4884]: I1202 01:52:27.702504 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-557b87d65-9tj59" event={"ID":"4f2e27ae-6dcb-4d32-b65b-f7d96a026d25","Type":"ContainerStarted","Data":"d43a6d31e442eb047a9c911191992af0f779e2252417279044d32882b4e3d428"} Dec 02 01:52:27 crc kubenswrapper[4884]: I1202 01:52:27.703658 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-557b87d65-9tj59" Dec 02 01:52:27 crc kubenswrapper[4884]: I1202 01:52:27.751736 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-557b87d65-9tj59" podStartSLOduration=2.044710037 podStartE2EDuration="5.751712792s" podCreationTimestamp="2025-12-02 01:52:22 +0000 UTC" firstStartedPulling="2025-12-02 01:52:23.336315849 +0000 UTC m=+960.012152733" lastFinishedPulling="2025-12-02 01:52:27.043318584 +0000 UTC m=+963.719155488" observedRunningTime="2025-12-02 01:52:27.746906352 +0000 UTC m=+964.422743326" watchObservedRunningTime="2025-12-02 01:52:27.751712792 +0000 UTC m=+964.427549716" Dec 02 01:52:32 crc kubenswrapper[4884]: I1202 01:52:32.814433 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-557b87d65-9tj59" Dec 02 01:52:46 crc kubenswrapper[4884]: I1202 01:52:46.971906 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:52:46 crc kubenswrapper[4884]: I1202 01:52:46.972331 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.003839 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-fspk2"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.005129 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fspk2" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.007082 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-jjjzr" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.044289 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-fspk2"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.051382 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bv5r\" (UniqueName: \"kubernetes.io/projected/50955276-266a-4e3d-9537-0efccdd1028d-kube-api-access-8bv5r\") pod \"barbican-operator-controller-manager-7d9dfd778-fspk2\" (UID: \"50955276-266a-4e3d-9537-0efccdd1028d\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fspk2" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.053855 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-7tjcj"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.059785 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7tjcj" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.071623 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-7rb9z" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.071801 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.072905 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.076243 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-hmf8t" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.076394 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-h8944"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.077792 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h8944" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.082148 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-4p2mg" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.082349 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-7tjcj"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.087777 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.095057 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4pvr8"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.096060 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4pvr8" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.100070 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-h8944"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.103467 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-7b78f" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.109260 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-j49bt"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.110668 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-j49bt" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.119343 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-7hbrp" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.121799 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4pvr8"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.125622 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-j49bt"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.137565 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.141524 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.144615 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.144859 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-6sh8r" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.146426 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.152249 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stc8k\" (UniqueName: \"kubernetes.io/projected/4cbce330-7938-453f-989a-c29213cf9118-kube-api-access-stc8k\") pod \"horizon-operator-controller-manager-68c6d99b8f-j49bt\" (UID: \"4cbce330-7938-453f-989a-c29213cf9118\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-j49bt" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.152289 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bv5r\" (UniqueName: \"kubernetes.io/projected/50955276-266a-4e3d-9537-0efccdd1028d-kube-api-access-8bv5r\") pod \"barbican-operator-controller-manager-7d9dfd778-fspk2\" (UID: \"50955276-266a-4e3d-9537-0efccdd1028d\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fspk2" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.152310 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvwtq\" (UniqueName: \"kubernetes.io/projected/6c738783-6d27-48e8-8f8e-a34f0f0ecef8-kube-api-access-wvwtq\") pod \"cinder-operator-controller-manager-859b6ccc6-7tjcj\" (UID: \"6c738783-6d27-48e8-8f8e-a34f0f0ecef8\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7tjcj" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.152337 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnrsd\" (UniqueName: \"kubernetes.io/projected/15629082-5a45-439e-b98e-57083adea72c-kube-api-access-rnrsd\") pod \"glance-operator-controller-manager-668d9c48b9-qpg2b\" (UID: \"15629082-5a45-439e-b98e-57083adea72c\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.152376 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zgrf\" (UniqueName: \"kubernetes.io/projected/68a5e45a-0241-4dad-97d5-3855e1128b01-kube-api-access-4zgrf\") pod \"heat-operator-controller-manager-5f64f6f8bb-4pvr8\" (UID: \"68a5e45a-0241-4dad-97d5-3855e1128b01\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4pvr8" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.152404 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2djtm\" (UniqueName: \"kubernetes.io/projected/ccee2404-f917-4dc3-bd71-be8c8aaf5362-kube-api-access-2djtm\") pod \"designate-operator-controller-manager-78b4bc895b-h8944\" (UID: \"ccee2404-f917-4dc3-bd71-be8c8aaf5362\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h8944" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.154793 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-kszl7"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.155862 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kszl7" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.158092 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-m47lf" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.175177 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-kszl7"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.184796 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-kmj5d"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.197034 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-kmj5d" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.198488 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bv5r\" (UniqueName: \"kubernetes.io/projected/50955276-266a-4e3d-9537-0efccdd1028d-kube-api-access-8bv5r\") pod \"barbican-operator-controller-manager-7d9dfd778-fspk2\" (UID: \"50955276-266a-4e3d-9537-0efccdd1028d\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fspk2" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.198989 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-v6f2g" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.256643 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-kmj5d"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.303387 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2djtm\" (UniqueName: \"kubernetes.io/projected/ccee2404-f917-4dc3-bd71-be8c8aaf5362-kube-api-access-2djtm\") pod \"designate-operator-controller-manager-78b4bc895b-h8944\" (UID: \"ccee2404-f917-4dc3-bd71-be8c8aaf5362\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h8944" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.303533 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stc8k\" (UniqueName: \"kubernetes.io/projected/4cbce330-7938-453f-989a-c29213cf9118-kube-api-access-stc8k\") pod \"horizon-operator-controller-manager-68c6d99b8f-j49bt\" (UID: \"4cbce330-7938-453f-989a-c29213cf9118\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-j49bt" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.303562 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvwtq\" (UniqueName: \"kubernetes.io/projected/6c738783-6d27-48e8-8f8e-a34f0f0ecef8-kube-api-access-wvwtq\") pod \"cinder-operator-controller-manager-859b6ccc6-7tjcj\" (UID: \"6c738783-6d27-48e8-8f8e-a34f0f0ecef8\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7tjcj" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.303586 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsvwg\" (UniqueName: \"kubernetes.io/projected/f3bc5025-5eac-4f60-a65c-c1abda8e6367-kube-api-access-qsvwg\") pod \"ironic-operator-controller-manager-6c548fd776-kszl7\" (UID: \"f3bc5025-5eac-4f60-a65c-c1abda8e6367\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kszl7" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.303626 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnrsd\" (UniqueName: \"kubernetes.io/projected/15629082-5a45-439e-b98e-57083adea72c-kube-api-access-rnrsd\") pod \"glance-operator-controller-manager-668d9c48b9-qpg2b\" (UID: \"15629082-5a45-439e-b98e-57083adea72c\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.303706 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zgrf\" (UniqueName: \"kubernetes.io/projected/68a5e45a-0241-4dad-97d5-3855e1128b01-kube-api-access-4zgrf\") pod \"heat-operator-controller-manager-5f64f6f8bb-4pvr8\" (UID: \"68a5e45a-0241-4dad-97d5-3855e1128b01\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4pvr8" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.303767 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert\") pod \"infra-operator-controller-manager-57548d458d-nhlqg\" (UID: \"6770b6e2-f0dc-4a64-96c7-3d0854fadf3d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.303785 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9zjs\" (UniqueName: \"kubernetes.io/projected/d7fbdcb3-5250-4b03-b182-47506f9e0c50-kube-api-access-p9zjs\") pod \"keystone-operator-controller-manager-546d4bdf48-kmj5d\" (UID: \"d7fbdcb3-5250-4b03-b182-47506f9e0c50\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-kmj5d" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.303808 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8d4n\" (UniqueName: \"kubernetes.io/projected/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-kube-api-access-k8d4n\") pod \"infra-operator-controller-manager-57548d458d-nhlqg\" (UID: \"6770b6e2-f0dc-4a64-96c7-3d0854fadf3d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.305392 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.314057 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.317655 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.322167 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-ghvt5" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.337117 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fspk2" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.360515 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnrsd\" (UniqueName: \"kubernetes.io/projected/15629082-5a45-439e-b98e-57083adea72c-kube-api-access-rnrsd\") pod \"glance-operator-controller-manager-668d9c48b9-qpg2b\" (UID: \"15629082-5a45-439e-b98e-57083adea72c\") " pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.362519 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stc8k\" (UniqueName: \"kubernetes.io/projected/4cbce330-7938-453f-989a-c29213cf9118-kube-api-access-stc8k\") pod \"horizon-operator-controller-manager-68c6d99b8f-j49bt\" (UID: \"4cbce330-7938-453f-989a-c29213cf9118\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-j49bt" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.362648 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvwtq\" (UniqueName: \"kubernetes.io/projected/6c738783-6d27-48e8-8f8e-a34f0f0ecef8-kube-api-access-wvwtq\") pod \"cinder-operator-controller-manager-859b6ccc6-7tjcj\" (UID: \"6c738783-6d27-48e8-8f8e-a34f0f0ecef8\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7tjcj" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.366468 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-d2df5"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.367606 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-d2df5" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.372337 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-kmp4s" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.376492 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2djtm\" (UniqueName: \"kubernetes.io/projected/ccee2404-f917-4dc3-bd71-be8c8aaf5362-kube-api-access-2djtm\") pod \"designate-operator-controller-manager-78b4bc895b-h8944\" (UID: \"ccee2404-f917-4dc3-bd71-be8c8aaf5362\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h8944" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.394572 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zgrf\" (UniqueName: \"kubernetes.io/projected/68a5e45a-0241-4dad-97d5-3855e1128b01-kube-api-access-4zgrf\") pod \"heat-operator-controller-manager-5f64f6f8bb-4pvr8\" (UID: \"68a5e45a-0241-4dad-97d5-3855e1128b01\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4pvr8" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.406422 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t7r7\" (UniqueName: \"kubernetes.io/projected/8b6aa6b7-93b2-4ba1-9a01-10c47d21df36-kube-api-access-9t7r7\") pod \"manila-operator-controller-manager-6546668bfd-dm29n\" (UID: \"8b6aa6b7-93b2-4ba1-9a01-10c47d21df36\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.406479 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsvwg\" (UniqueName: \"kubernetes.io/projected/f3bc5025-5eac-4f60-a65c-c1abda8e6367-kube-api-access-qsvwg\") pod \"ironic-operator-controller-manager-6c548fd776-kszl7\" (UID: \"f3bc5025-5eac-4f60-a65c-c1abda8e6367\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kszl7" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.406548 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert\") pod \"infra-operator-controller-manager-57548d458d-nhlqg\" (UID: \"6770b6e2-f0dc-4a64-96c7-3d0854fadf3d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.406567 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9zjs\" (UniqueName: \"kubernetes.io/projected/d7fbdcb3-5250-4b03-b182-47506f9e0c50-kube-api-access-p9zjs\") pod \"keystone-operator-controller-manager-546d4bdf48-kmj5d\" (UID: \"d7fbdcb3-5250-4b03-b182-47506f9e0c50\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-kmj5d" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.406587 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8d4n\" (UniqueName: \"kubernetes.io/projected/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-kube-api-access-k8d4n\") pod \"infra-operator-controller-manager-57548d458d-nhlqg\" (UID: \"6770b6e2-f0dc-4a64-96c7-3d0854fadf3d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.406701 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7tjcj" Dec 02 01:52:53 crc kubenswrapper[4884]: E1202 01:52:53.407146 4884 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 01:52:53 crc kubenswrapper[4884]: E1202 01:52:53.407186 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert podName:6770b6e2-f0dc-4a64-96c7-3d0854fadf3d nodeName:}" failed. No retries permitted until 2025-12-02 01:52:53.907171373 +0000 UTC m=+990.583008257 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert") pod "infra-operator-controller-manager-57548d458d-nhlqg" (UID: "6770b6e2-f0dc-4a64-96c7-3d0854fadf3d") : secret "infra-operator-webhook-server-cert" not found Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.422594 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.423918 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.427042 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.440427 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8d4n\" (UniqueName: \"kubernetes.io/projected/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-kube-api-access-k8d4n\") pod \"infra-operator-controller-manager-57548d458d-nhlqg\" (UID: \"6770b6e2-f0dc-4a64-96c7-3d0854fadf3d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.440715 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h8944" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.448805 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-d2df5"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.451160 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-df65h" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.459652 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4pvr8" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.461323 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsvwg\" (UniqueName: \"kubernetes.io/projected/f3bc5025-5eac-4f60-a65c-c1abda8e6367-kube-api-access-qsvwg\") pod \"ironic-operator-controller-manager-6c548fd776-kszl7\" (UID: \"f3bc5025-5eac-4f60-a65c-c1abda8e6367\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kszl7" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.462477 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.474707 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9zjs\" (UniqueName: \"kubernetes.io/projected/d7fbdcb3-5250-4b03-b182-47506f9e0c50-kube-api-access-p9zjs\") pod \"keystone-operator-controller-manager-546d4bdf48-kmj5d\" (UID: \"d7fbdcb3-5250-4b03-b182-47506f9e0c50\") " pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-kmj5d" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.475730 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.476847 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.482929 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-2pm88" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.486083 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-j49bt" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.508564 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6zq8\" (UniqueName: \"kubernetes.io/projected/8e4db64b-9154-48cd-bd3c-463038b4dd93-kube-api-access-l6zq8\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-r8hgf\" (UID: \"8e4db64b-9154-48cd-bd3c-463038b4dd93\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.508632 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t7r7\" (UniqueName: \"kubernetes.io/projected/8b6aa6b7-93b2-4ba1-9a01-10c47d21df36-kube-api-access-9t7r7\") pod \"manila-operator-controller-manager-6546668bfd-dm29n\" (UID: \"8b6aa6b7-93b2-4ba1-9a01-10c47d21df36\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.508659 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwbpr\" (UniqueName: \"kubernetes.io/projected/17378750-b1cf-4515-81f4-638404739f0f-kube-api-access-wwbpr\") pod \"mariadb-operator-controller-manager-56bbcc9d85-d2df5\" (UID: \"17378750-b1cf-4515-81f4-638404739f0f\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-d2df5" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.518953 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.537174 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t7r7\" (UniqueName: \"kubernetes.io/projected/8b6aa6b7-93b2-4ba1-9a01-10c47d21df36-kube-api-access-9t7r7\") pod \"manila-operator-controller-manager-6546668bfd-dm29n\" (UID: \"8b6aa6b7-93b2-4ba1-9a01-10c47d21df36\") " pod="openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.544924 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-vz95x"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.546336 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vz95x" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.560359 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-vz95x"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.570026 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kszl7" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.571650 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-rk6ds" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.579517 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.580614 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.591068 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.591360 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-6wppl" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.592521 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-kf879"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.593549 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-kf879" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.606642 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-nzwwd"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.607728 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nzwwd" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.612249 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-j48gg" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.612480 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-btbvj" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.613145 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbrlh\" (UniqueName: \"kubernetes.io/projected/3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f-kube-api-access-zbrlh\") pod \"nova-operator-controller-manager-697bc559fc-w5xb8\" (UID: \"3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.613178 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6zq8\" (UniqueName: \"kubernetes.io/projected/8e4db64b-9154-48cd-bd3c-463038b4dd93-kube-api-access-l6zq8\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-r8hgf\" (UID: \"8e4db64b-9154-48cd-bd3c-463038b4dd93\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.613239 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwbpr\" (UniqueName: \"kubernetes.io/projected/17378750-b1cf-4515-81f4-638404739f0f-kube-api-access-wwbpr\") pod \"mariadb-operator-controller-manager-56bbcc9d85-d2df5\" (UID: \"17378750-b1cf-4515-81f4-638404739f0f\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-d2df5" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.613279 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq4dg\" (UniqueName: \"kubernetes.io/projected/58dc93d7-c6f5-4927-9a80-887c561cbe3b-kube-api-access-zq4dg\") pod \"octavia-operator-controller-manager-998648c74-vz95x\" (UID: \"58dc93d7-c6f5-4927-9a80-887c561cbe3b\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-vz95x" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.630881 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-kmj5d" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.650386 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6zq8\" (UniqueName: \"kubernetes.io/projected/8e4db64b-9154-48cd-bd3c-463038b4dd93-kube-api-access-l6zq8\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-r8hgf\" (UID: \"8e4db64b-9154-48cd-bd3c-463038b4dd93\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.662434 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.670194 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.671575 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.676850 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwbpr\" (UniqueName: \"kubernetes.io/projected/17378750-b1cf-4515-81f4-638404739f0f-kube-api-access-wwbpr\") pod \"mariadb-operator-controller-manager-56bbcc9d85-d2df5\" (UID: \"17378750-b1cf-4515-81f4-638404739f0f\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-d2df5" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.680960 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-cbprc" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.687796 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.688894 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.695050 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-mq75h" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.714565 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckxkp\" (UniqueName: \"kubernetes.io/projected/edf5e61d-2bf9-4eb6-9363-88e76f4e9506-kube-api-access-ckxkp\") pod \"swift-operator-controller-manager-5f8c65bbfc-znmww\" (UID: \"edf5e61d-2bf9-4eb6-9363-88e76f4e9506\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.714852 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq4dg\" (UniqueName: \"kubernetes.io/projected/58dc93d7-c6f5-4927-9a80-887c561cbe3b-kube-api-access-zq4dg\") pod \"octavia-operator-controller-manager-998648c74-vz95x\" (UID: \"58dc93d7-c6f5-4927-9a80-887c561cbe3b\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-vz95x" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.714876 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd456f66\" (UID: \"2fca9668-22fe-4ea5-a438-f72c6e9c5294\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.714894 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62bzq\" (UniqueName: \"kubernetes.io/projected/2fca9668-22fe-4ea5-a438-f72c6e9c5294-kube-api-access-62bzq\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd456f66\" (UID: \"2fca9668-22fe-4ea5-a438-f72c6e9c5294\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.714945 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbrlh\" (UniqueName: \"kubernetes.io/projected/3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f-kube-api-access-zbrlh\") pod \"nova-operator-controller-manager-697bc559fc-w5xb8\" (UID: \"3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.714961 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fckjj\" (UniqueName: \"kubernetes.io/projected/c9178c86-b9ac-4d5b-8145-0cc5f0e79157-kube-api-access-fckjj\") pod \"ovn-operator-controller-manager-b6456fdb6-kf879\" (UID: \"c9178c86-b9ac-4d5b-8145-0cc5f0e79157\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-kf879" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.714982 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np5cf\" (UniqueName: \"kubernetes.io/projected/dd662773-0d4d-4682-9fdb-b79667483d76-kube-api-access-np5cf\") pod \"placement-operator-controller-manager-78f8948974-nzwwd\" (UID: \"dd662773-0d4d-4682-9fdb-b79667483d76\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-nzwwd" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.732980 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-kf879"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.737799 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq4dg\" (UniqueName: \"kubernetes.io/projected/58dc93d7-c6f5-4927-9a80-887c561cbe3b-kube-api-access-zq4dg\") pod \"octavia-operator-controller-manager-998648c74-vz95x\" (UID: \"58dc93d7-c6f5-4927-9a80-887c561cbe3b\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-vz95x" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.744368 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbrlh\" (UniqueName: \"kubernetes.io/projected/3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f-kube-api-access-zbrlh\") pod \"nova-operator-controller-manager-697bc559fc-w5xb8\" (UID: \"3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.747284 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.768220 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-nzwwd"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.821639 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckxkp\" (UniqueName: \"kubernetes.io/projected/edf5e61d-2bf9-4eb6-9363-88e76f4e9506-kube-api-access-ckxkp\") pod \"swift-operator-controller-manager-5f8c65bbfc-znmww\" (UID: \"edf5e61d-2bf9-4eb6-9363-88e76f4e9506\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.821701 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j5qj\" (UniqueName: \"kubernetes.io/projected/c7c9137a-0bc9-41e0-9dc0-f5976861a3b2-kube-api-access-4j5qj\") pod \"telemetry-operator-controller-manager-76cc84c6bb-x5hdr\" (UID: \"c7c9137a-0bc9-41e0-9dc0-f5976861a3b2\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.821730 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62bzq\" (UniqueName: \"kubernetes.io/projected/2fca9668-22fe-4ea5-a438-f72c6e9c5294-kube-api-access-62bzq\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd456f66\" (UID: \"2fca9668-22fe-4ea5-a438-f72c6e9c5294\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.821776 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd456f66\" (UID: \"2fca9668-22fe-4ea5-a438-f72c6e9c5294\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" Dec 02 01:52:53 crc kubenswrapper[4884]: E1202 01:52:53.821927 4884 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.821926 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fckjj\" (UniqueName: \"kubernetes.io/projected/c9178c86-b9ac-4d5b-8145-0cc5f0e79157-kube-api-access-fckjj\") pod \"ovn-operator-controller-manager-b6456fdb6-kf879\" (UID: \"c9178c86-b9ac-4d5b-8145-0cc5f0e79157\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-kf879" Dec 02 01:52:53 crc kubenswrapper[4884]: E1202 01:52:53.821999 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert podName:2fca9668-22fe-4ea5-a438-f72c6e9c5294 nodeName:}" failed. No retries permitted until 2025-12-02 01:52:54.321969492 +0000 UTC m=+990.997806376 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd456f66" (UID: "2fca9668-22fe-4ea5-a438-f72c6e9c5294") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.822032 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np5cf\" (UniqueName: \"kubernetes.io/projected/dd662773-0d4d-4682-9fdb-b79667483d76-kube-api-access-np5cf\") pod \"placement-operator-controller-manager-78f8948974-nzwwd\" (UID: \"dd662773-0d4d-4682-9fdb-b79667483d76\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-nzwwd" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.829844 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.854595 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.857694 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.863993 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.866273 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.871839 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-ldp9h" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.873850 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-pr8hc" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.882611 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-d2df5" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.893199 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.893518 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.898492 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np5cf\" (UniqueName: \"kubernetes.io/projected/dd662773-0d4d-4682-9fdb-b79667483d76-kube-api-access-np5cf\") pod \"placement-operator-controller-manager-78f8948974-nzwwd\" (UID: \"dd662773-0d4d-4682-9fdb-b79667483d76\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-nzwwd" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.898821 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckxkp\" (UniqueName: \"kubernetes.io/projected/edf5e61d-2bf9-4eb6-9363-88e76f4e9506-kube-api-access-ckxkp\") pod \"swift-operator-controller-manager-5f8c65bbfc-znmww\" (UID: \"edf5e61d-2bf9-4eb6-9363-88e76f4e9506\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.898840 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62bzq\" (UniqueName: \"kubernetes.io/projected/2fca9668-22fe-4ea5-a438-f72c6e9c5294-kube-api-access-62bzq\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd456f66\" (UID: \"2fca9668-22fe-4ea5-a438-f72c6e9c5294\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.898872 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fckjj\" (UniqueName: \"kubernetes.io/projected/c9178c86-b9ac-4d5b-8145-0cc5f0e79157-kube-api-access-fckjj\") pod \"ovn-operator-controller-manager-b6456fdb6-kf879\" (UID: \"c9178c86-b9ac-4d5b-8145-0cc5f0e79157\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-kf879" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.900370 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.913299 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.919014 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.925468 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j5qj\" (UniqueName: \"kubernetes.io/projected/c7c9137a-0bc9-41e0-9dc0-f5976861a3b2-kube-api-access-4j5qj\") pod \"telemetry-operator-controller-manager-76cc84c6bb-x5hdr\" (UID: \"c7c9137a-0bc9-41e0-9dc0-f5976861a3b2\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.925508 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwq6n\" (UniqueName: \"kubernetes.io/projected/264f2ec8-0c0a-4fce-8ed3-eb8fafae0621-kube-api-access-nwq6n\") pod \"watcher-operator-controller-manager-84f44f5d76-fz96n\" (UID: \"264f2ec8-0c0a-4fce-8ed3-eb8fafae0621\") " pod="openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.925564 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert\") pod \"infra-operator-controller-manager-57548d458d-nhlqg\" (UID: \"6770b6e2-f0dc-4a64-96c7-3d0854fadf3d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.925634 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n5g2\" (UniqueName: \"kubernetes.io/projected/4446fb10-1203-4063-affd-bc81cbfa4238-kube-api-access-7n5g2\") pod \"test-operator-controller-manager-5854674fcc-nnfmw\" (UID: \"4446fb10-1203-4063-affd-bc81cbfa4238\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw" Dec 02 01:52:53 crc kubenswrapper[4884]: E1202 01:52:53.925803 4884 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 01:52:53 crc kubenswrapper[4884]: E1202 01:52:53.925871 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert podName:6770b6e2-f0dc-4a64-96c7-3d0854fadf3d nodeName:}" failed. No retries permitted until 2025-12-02 01:52:54.925857648 +0000 UTC m=+991.601694532 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert") pod "infra-operator-controller-manager-57548d458d-nhlqg" (UID: "6770b6e2-f0dc-4a64-96c7-3d0854fadf3d") : secret "infra-operator-webhook-server-cert" not found Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.940555 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.941558 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.942687 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vz95x" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.948530 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.948679 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-stmn2" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.948732 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.952531 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.956303 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.957614 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.959769 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-jlwsz" Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.960168 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8"] Dec 02 01:52:53 crc kubenswrapper[4884]: I1202 01:52:53.964008 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j5qj\" (UniqueName: \"kubernetes.io/projected/c7c9137a-0bc9-41e0-9dc0-f5976861a3b2-kube-api-access-4j5qj\") pod \"telemetry-operator-controller-manager-76cc84c6bb-x5hdr\" (UID: \"c7c9137a-0bc9-41e0-9dc0-f5976861a3b2\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.008443 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-kf879" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.035397 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nzwwd" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.036155 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.036201 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n5g2\" (UniqueName: \"kubernetes.io/projected/4446fb10-1203-4063-affd-bc81cbfa4238-kube-api-access-7n5g2\") pod \"test-operator-controller-manager-5854674fcc-nnfmw\" (UID: \"4446fb10-1203-4063-affd-bc81cbfa4238\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.036251 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwq6n\" (UniqueName: \"kubernetes.io/projected/264f2ec8-0c0a-4fce-8ed3-eb8fafae0621-kube-api-access-nwq6n\") pod \"watcher-operator-controller-manager-84f44f5d76-fz96n\" (UID: \"264f2ec8-0c0a-4fce-8ed3-eb8fafae0621\") " pod="openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.036273 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.036299 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq9ks\" (UniqueName: \"kubernetes.io/projected/d7c0cc8c-1763-443f-a2a0-5d76d093b019-kube-api-access-jq9ks\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qjjv8\" (UID: \"d7c0cc8c-1763-443f-a2a0-5d76d093b019\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.036336 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rz8b\" (UniqueName: \"kubernetes.io/projected/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-kube-api-access-9rz8b\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.054433 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwq6n\" (UniqueName: \"kubernetes.io/projected/264f2ec8-0c0a-4fce-8ed3-eb8fafae0621-kube-api-access-nwq6n\") pod \"watcher-operator-controller-manager-84f44f5d76-fz96n\" (UID: \"264f2ec8-0c0a-4fce-8ed3-eb8fafae0621\") " pod="openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.065512 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.069707 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n5g2\" (UniqueName: \"kubernetes.io/projected/4446fb10-1203-4063-affd-bc81cbfa4238-kube-api-access-7n5g2\") pod \"test-operator-controller-manager-5854674fcc-nnfmw\" (UID: \"4446fb10-1203-4063-affd-bc81cbfa4238\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.138776 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.139634 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.139695 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.139716 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq9ks\" (UniqueName: \"kubernetes.io/projected/d7c0cc8c-1763-443f-a2a0-5d76d093b019-kube-api-access-jq9ks\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qjjv8\" (UID: \"d7c0cc8c-1763-443f-a2a0-5d76d093b019\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.139767 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rz8b\" (UniqueName: \"kubernetes.io/projected/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-kube-api-access-9rz8b\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:52:54 crc kubenswrapper[4884]: E1202 01:52:54.140096 4884 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 01:52:54 crc kubenswrapper[4884]: E1202 01:52:54.140144 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs podName:130cad20-8dfc-4f3d-a00f-d0d5d066ef54 nodeName:}" failed. No retries permitted until 2025-12-02 01:52:54.640130464 +0000 UTC m=+991.315967348 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs") pod "openstack-operator-controller-manager-668dd49799-k2fql" (UID: "130cad20-8dfc-4f3d-a00f-d0d5d066ef54") : secret "metrics-server-cert" not found Dec 02 01:52:54 crc kubenswrapper[4884]: E1202 01:52:54.140291 4884 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 01:52:54 crc kubenswrapper[4884]: E1202 01:52:54.140377 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs podName:130cad20-8dfc-4f3d-a00f-d0d5d066ef54 nodeName:}" failed. No retries permitted until 2025-12-02 01:52:54.640351489 +0000 UTC m=+991.316188373 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs") pod "openstack-operator-controller-manager-668dd49799-k2fql" (UID: "130cad20-8dfc-4f3d-a00f-d0d5d066ef54") : secret "webhook-server-cert" not found Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.155320 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rz8b\" (UniqueName: \"kubernetes.io/projected/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-kube-api-access-9rz8b\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.160375 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq9ks\" (UniqueName: \"kubernetes.io/projected/d7c0cc8c-1763-443f-a2a0-5d76d093b019-kube-api-access-jq9ks\") pod \"rabbitmq-cluster-operator-manager-668c99d594-qjjv8\" (UID: \"d7c0cc8c-1763-443f-a2a0-5d76d093b019\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.277369 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.306694 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.342165 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd456f66\" (UID: \"2fca9668-22fe-4ea5-a438-f72c6e9c5294\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" Dec 02 01:52:54 crc kubenswrapper[4884]: E1202 01:52:54.342461 4884 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 01:52:54 crc kubenswrapper[4884]: E1202 01:52:54.342593 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert podName:2fca9668-22fe-4ea5-a438-f72c6e9c5294 nodeName:}" failed. No retries permitted until 2025-12-02 01:52:55.342560334 +0000 UTC m=+992.018397218 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd456f66" (UID: "2fca9668-22fe-4ea5-a438-f72c6e9c5294") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.357193 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8" Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.480825 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-fspk2"] Dec 02 01:52:54 crc kubenswrapper[4884]: W1202 01:52:54.512970 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50955276_266a_4e3d_9537_0efccdd1028d.slice/crio-a0b988184f4a2753c3190697c6a4471abcb9eb78e494f1afe28048de761f0193 WatchSource:0}: Error finding container a0b988184f4a2753c3190697c6a4471abcb9eb78e494f1afe28048de761f0193: Status 404 returned error can't find the container with id a0b988184f4a2753c3190697c6a4471abcb9eb78e494f1afe28048de761f0193 Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.525979 4884 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.647993 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:52:54 crc kubenswrapper[4884]: E1202 01:52:54.648268 4884 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.648317 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:52:54 crc kubenswrapper[4884]: E1202 01:52:54.648386 4884 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 01:52:54 crc kubenswrapper[4884]: E1202 01:52:54.648444 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs podName:130cad20-8dfc-4f3d-a00f-d0d5d066ef54 nodeName:}" failed. No retries permitted until 2025-12-02 01:52:55.648424779 +0000 UTC m=+992.324261683 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs") pod "openstack-operator-controller-manager-668dd49799-k2fql" (UID: "130cad20-8dfc-4f3d-a00f-d0d5d066ef54") : secret "webhook-server-cert" not found Dec 02 01:52:54 crc kubenswrapper[4884]: E1202 01:52:54.648859 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs podName:130cad20-8dfc-4f3d-a00f-d0d5d066ef54 nodeName:}" failed. No retries permitted until 2025-12-02 01:52:55.64884483 +0000 UTC m=+992.324681734 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs") pod "openstack-operator-controller-manager-668dd49799-k2fql" (UID: "130cad20-8dfc-4f3d-a00f-d0d5d066ef54") : secret "metrics-server-cert" not found Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.804914 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-7tjcj"] Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.816170 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-h8944"] Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.821894 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b"] Dec 02 01:52:54 crc kubenswrapper[4884]: W1202 01:52:54.824443 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccee2404_f917_4dc3_bd71_be8c8aaf5362.slice/crio-dac228b0a7bba36cc6bb107f049fb770582c80ef3c1504a329e0ade0f7230fa2 WatchSource:0}: Error finding container dac228b0a7bba36cc6bb107f049fb770582c80ef3c1504a329e0ade0f7230fa2: Status 404 returned error can't find the container with id dac228b0a7bba36cc6bb107f049fb770582c80ef3c1504a329e0ade0f7230fa2 Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.875945 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-j49bt"] Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.913361 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b" event={"ID":"15629082-5a45-439e-b98e-57083adea72c","Type":"ContainerStarted","Data":"dfd677ae4ce40453abdb8e8b4bc8f197074a7e5dff0859d378cc73158846ae4f"} Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.914574 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h8944" event={"ID":"ccee2404-f917-4dc3-bd71-be8c8aaf5362","Type":"ContainerStarted","Data":"dac228b0a7bba36cc6bb107f049fb770582c80ef3c1504a329e0ade0f7230fa2"} Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.915549 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fspk2" event={"ID":"50955276-266a-4e3d-9537-0efccdd1028d","Type":"ContainerStarted","Data":"a0b988184f4a2753c3190697c6a4471abcb9eb78e494f1afe28048de761f0193"} Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.916612 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-j49bt" event={"ID":"4cbce330-7938-453f-989a-c29213cf9118","Type":"ContainerStarted","Data":"42ab362376d2679fccd749592c4180907ac7b617ec5792d6ecc40ebfb40cc2a5"} Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.917441 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7tjcj" event={"ID":"6c738783-6d27-48e8-8f8e-a34f0f0ecef8","Type":"ContainerStarted","Data":"09a4ae35c3900b86ca352171451c4ed021fb413e364d9eede1d805e7a0b03887"} Dec 02 01:52:54 crc kubenswrapper[4884]: I1202 01:52:54.957858 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert\") pod \"infra-operator-controller-manager-57548d458d-nhlqg\" (UID: \"6770b6e2-f0dc-4a64-96c7-3d0854fadf3d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" Dec 02 01:52:54 crc kubenswrapper[4884]: E1202 01:52:54.958153 4884 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 01:52:54 crc kubenswrapper[4884]: E1202 01:52:54.958285 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert podName:6770b6e2-f0dc-4a64-96c7-3d0854fadf3d nodeName:}" failed. No retries permitted until 2025-12-02 01:52:56.958252874 +0000 UTC m=+993.634089948 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert") pod "infra-operator-controller-manager-57548d458d-nhlqg" (UID: "6770b6e2-f0dc-4a64-96c7-3d0854fadf3d") : secret "infra-operator-webhook-server-cert" not found Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.091038 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-d2df5"] Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.100313 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-546d4bdf48-kmj5d"] Dec 02 01:52:55 crc kubenswrapper[4884]: W1202 01:52:55.108080 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17378750_b1cf_4515_81f4_638404739f0f.slice/crio-9cc2c77489788bf06a3f51e2ebc008db6354a106a8db253099fba77c41cc4b8e WatchSource:0}: Error finding container 9cc2c77489788bf06a3f51e2ebc008db6354a106a8db253099fba77c41cc4b8e: Status 404 returned error can't find the container with id 9cc2c77489788bf06a3f51e2ebc008db6354a106a8db253099fba77c41cc4b8e Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.110248 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4pvr8"] Dec 02 01:52:55 crc kubenswrapper[4884]: W1202 01:52:55.113412 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7fbdcb3_5250_4b03_b182_47506f9e0c50.slice/crio-4bd39d0562ef0036d75e124c2a847464c0cd45549444b6917c71c623aad9d6ce WatchSource:0}: Error finding container 4bd39d0562ef0036d75e124c2a847464c0cd45549444b6917c71c623aad9d6ce: Status 404 returned error can't find the container with id 4bd39d0562ef0036d75e124c2a847464c0cd45549444b6917c71c623aad9d6ce Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.148042 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-kszl7"] Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.202499 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n"] Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.215129 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-vz95x"] Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.220407 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ckxkp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-znmww_openstack-operators(edf5e61d-2bf9-4eb6-9363-88e76f4e9506): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.223593 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ckxkp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-znmww_openstack-operators(edf5e61d-2bf9-4eb6-9363-88e76f4e9506): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.224868 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zbrlh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-w5xb8_openstack-operators(3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.224956 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww" podUID="edf5e61d-2bf9-4eb6-9363-88e76f4e9506" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.227481 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zbrlh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-w5xb8_openstack-operators(3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.228917 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8" podUID="3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f" Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.230668 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-nzwwd"] Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.242944 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-kf879"] Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.246361 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.5:5001/openstack-k8s-operators/watcher-operator:a03c42dd21a97c49a51c6a865b0d8e37fb31f961,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nwq6n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-84f44f5d76-fz96n_openstack-operators(264f2ec8-0c0a-4fce-8ed3-eb8fafae0621): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 01:52:55 crc kubenswrapper[4884]: W1202 01:52:55.247326 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7c9137a_0bc9_41e0_9dc0_f5976861a3b2.slice/crio-9e6a146c5bb089de9baed106d3211407f9a8f6998bb4116efb7c26219b413a38 WatchSource:0}: Error finding container 9e6a146c5bb089de9baed106d3211407f9a8f6998bb4116efb7c26219b413a38: Status 404 returned error can't find the container with id 9e6a146c5bb089de9baed106d3211407f9a8f6998bb4116efb7c26219b413a38 Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.247915 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l6zq8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-r8hgf_openstack-operators(8e4db64b-9154-48cd-bd3c-463038b4dd93): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.248137 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7n5g2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-nnfmw_openstack-operators(4446fb10-1203-4063-affd-bc81cbfa4238): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.248603 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nwq6n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-84f44f5d76-fz96n_openstack-operators(264f2ec8-0c0a-4fce-8ed3-eb8fafae0621): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.250056 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n" podUID="264f2ec8-0c0a-4fce-8ed3-eb8fafae0621" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.250653 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7n5g2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-nnfmw_openstack-operators(4446fb10-1203-4063-affd-bc81cbfa4238): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.250733 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l6zq8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-r8hgf_openstack-operators(8e4db64b-9154-48cd-bd3c-463038b4dd93): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.251858 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf" podUID="8e4db64b-9154-48cd-bd3c-463038b4dd93" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.251905 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw" podUID="4446fb10-1203-4063-affd-bc81cbfa4238" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.252492 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jq9ks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-qjjv8_openstack-operators(d7c0cc8c-1763-443f-a2a0-5d76d093b019): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.252480 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4j5qj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-x5hdr_openstack-operators(c7c9137a-0bc9-41e0-9dc0-f5976861a3b2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.254092 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8" podUID="d7c0cc8c-1763-443f-a2a0-5d76d093b019" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.254688 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4j5qj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-x5hdr_openstack-operators(c7c9137a-0bc9-41e0-9dc0-f5976861a3b2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.256116 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" podUID="c7c9137a-0bc9-41e0-9dc0-f5976861a3b2" Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.256377 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww"] Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.261567 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf"] Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.266002 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8"] Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.272471 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n"] Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.275891 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw"] Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.280918 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr"] Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.285949 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8"] Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.399735 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd456f66\" (UID: \"2fca9668-22fe-4ea5-a438-f72c6e9c5294\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.399998 4884 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.400145 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert podName:2fca9668-22fe-4ea5-a438-f72c6e9c5294 nodeName:}" failed. No retries permitted until 2025-12-02 01:52:57.400068027 +0000 UTC m=+994.075904911 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd456f66" (UID: "2fca9668-22fe-4ea5-a438-f72c6e9c5294") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.705028 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.705448 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.706151 4884 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.706217 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs podName:130cad20-8dfc-4f3d-a00f-d0d5d066ef54 nodeName:}" failed. No retries permitted until 2025-12-02 01:52:57.706201398 +0000 UTC m=+994.382038282 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs") pod "openstack-operator-controller-manager-668dd49799-k2fql" (UID: "130cad20-8dfc-4f3d-a00f-d0d5d066ef54") : secret "webhook-server-cert" not found Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.706164 4884 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.706723 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs podName:130cad20-8dfc-4f3d-a00f-d0d5d066ef54 nodeName:}" failed. No retries permitted until 2025-12-02 01:52:57.706497545 +0000 UTC m=+994.382334549 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs") pod "openstack-operator-controller-manager-668dd49799-k2fql" (UID: "130cad20-8dfc-4f3d-a00f-d0d5d066ef54") : secret "metrics-server-cert" not found Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.929673 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw" event={"ID":"4446fb10-1203-4063-affd-bc81cbfa4238","Type":"ContainerStarted","Data":"1e21374440e1ef5233e8a71fa1be0d5c7ade8a3fc2e08b927c1799507488aa50"} Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.930951 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4pvr8" event={"ID":"68a5e45a-0241-4dad-97d5-3855e1128b01","Type":"ContainerStarted","Data":"7c9a1ae7ec1cecb11f45522ae4ddf2e119056f774d89a111f862ec19a54a9754"} Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.933371 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n" event={"ID":"8b6aa6b7-93b2-4ba1-9a01-10c47d21df36","Type":"ContainerStarted","Data":"84d7d9ac1fb182778c72f3cbb80d073f5108cb902e3a14c9a13a4e1afd76d366"} Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.935728 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw" podUID="4446fb10-1203-4063-affd-bc81cbfa4238" Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.936582 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" event={"ID":"c7c9137a-0bc9-41e0-9dc0-f5976861a3b2","Type":"ContainerStarted","Data":"9e6a146c5bb089de9baed106d3211407f9a8f6998bb4116efb7c26219b413a38"} Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.939831 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" podUID="c7c9137a-0bc9-41e0-9dc0-f5976861a3b2" Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.943524 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nzwwd" event={"ID":"dd662773-0d4d-4682-9fdb-b79667483d76","Type":"ContainerStarted","Data":"cd32ac295faab965545e29dfe48f50f7eacacafaaf7a5b500839b78243b2581e"} Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.956483 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kszl7" event={"ID":"f3bc5025-5eac-4f60-a65c-c1abda8e6367","Type":"ContainerStarted","Data":"58c3efad24bafdbce91434fd1f32eb1f0f2403f03bca918d9f5d928c6dd3f6ce"} Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.960335 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n" event={"ID":"264f2ec8-0c0a-4fce-8ed3-eb8fafae0621","Type":"ContainerStarted","Data":"d51a376945886491b5014ff0e4d8549b2310c08917d6cd430c5baa6c58c54c05"} Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.962878 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8" event={"ID":"d7c0cc8c-1763-443f-a2a0-5d76d093b019","Type":"ContainerStarted","Data":"a79e91e843fc3e87690c4d5f9e45458cf0c5f325174af226a9650a7a20e45dc4"} Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.965712 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8" podUID="d7c0cc8c-1763-443f-a2a0-5d76d093b019" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.966031 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.5:5001/openstack-k8s-operators/watcher-operator:a03c42dd21a97c49a51c6a865b0d8e37fb31f961\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n" podUID="264f2ec8-0c0a-4fce-8ed3-eb8fafae0621" Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.966819 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww" event={"ID":"edf5e61d-2bf9-4eb6-9363-88e76f4e9506","Type":"ContainerStarted","Data":"fafa4b898bf18f7fd7ab4e0b6b6f1249faec34847ac6461cade636ab26bfcb15"} Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.969398 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-kmj5d" event={"ID":"d7fbdcb3-5250-4b03-b182-47506f9e0c50","Type":"ContainerStarted","Data":"4bd39d0562ef0036d75e124c2a847464c0cd45549444b6917c71c623aad9d6ce"} Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.972329 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vz95x" event={"ID":"58dc93d7-c6f5-4927-9a80-887c561cbe3b","Type":"ContainerStarted","Data":"f0d07b003e8d811d620c63bb4ba120869561cb39796f025cadf8c335c39bacb3"} Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.972452 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww" podUID="edf5e61d-2bf9-4eb6-9363-88e76f4e9506" Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.976130 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-kf879" event={"ID":"c9178c86-b9ac-4d5b-8145-0cc5f0e79157","Type":"ContainerStarted","Data":"30f9315a25ea3a693eb99bd7914fadcfd735673bfcd333259dc2fefd94bd8657"} Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.978222 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-d2df5" event={"ID":"17378750-b1cf-4515-81f4-638404739f0f","Type":"ContainerStarted","Data":"9cc2c77489788bf06a3f51e2ebc008db6354a106a8db253099fba77c41cc4b8e"} Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.984900 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8" event={"ID":"3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f","Type":"ContainerStarted","Data":"7e490b566d7e21e956c377d6ce1238304932b7b8bef7f63c81e10abede13fe40"} Dec 02 01:52:55 crc kubenswrapper[4884]: I1202 01:52:55.987295 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf" event={"ID":"8e4db64b-9154-48cd-bd3c-463038b4dd93","Type":"ContainerStarted","Data":"7eeadc8539d80337bbbc64cb061679b702e39a9e7701ec0e18fb17a96359eb29"} Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.991638 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf" podUID="8e4db64b-9154-48cd-bd3c-463038b4dd93" Dec 02 01:52:55 crc kubenswrapper[4884]: E1202 01:52:55.993963 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8" podUID="3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f" Dec 02 01:52:57 crc kubenswrapper[4884]: E1202 01:52:57.001535 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8" podUID="d7c0cc8c-1763-443f-a2a0-5d76d093b019" Dec 02 01:52:57 crc kubenswrapper[4884]: E1202 01:52:57.002453 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww" podUID="edf5e61d-2bf9-4eb6-9363-88e76f4e9506" Dec 02 01:52:57 crc kubenswrapper[4884]: E1202 01:52:57.003700 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" podUID="c7c9137a-0bc9-41e0-9dc0-f5976861a3b2" Dec 02 01:52:57 crc kubenswrapper[4884]: E1202 01:52:57.003777 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf" podUID="8e4db64b-9154-48cd-bd3c-463038b4dd93" Dec 02 01:52:57 crc kubenswrapper[4884]: E1202 01:52:57.003852 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.5:5001/openstack-k8s-operators/watcher-operator:a03c42dd21a97c49a51c6a865b0d8e37fb31f961\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n" podUID="264f2ec8-0c0a-4fce-8ed3-eb8fafae0621" Dec 02 01:52:57 crc kubenswrapper[4884]: E1202 01:52:57.009663 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw" podUID="4446fb10-1203-4063-affd-bc81cbfa4238" Dec 02 01:52:57 crc kubenswrapper[4884]: E1202 01:52:57.014996 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8" podUID="3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f" Dec 02 01:52:57 crc kubenswrapper[4884]: I1202 01:52:57.026516 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert\") pod \"infra-operator-controller-manager-57548d458d-nhlqg\" (UID: \"6770b6e2-f0dc-4a64-96c7-3d0854fadf3d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" Dec 02 01:52:57 crc kubenswrapper[4884]: E1202 01:52:57.026662 4884 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 01:52:57 crc kubenswrapper[4884]: E1202 01:52:57.026708 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert podName:6770b6e2-f0dc-4a64-96c7-3d0854fadf3d nodeName:}" failed. No retries permitted until 2025-12-02 01:53:01.026694315 +0000 UTC m=+997.702531199 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert") pod "infra-operator-controller-manager-57548d458d-nhlqg" (UID: "6770b6e2-f0dc-4a64-96c7-3d0854fadf3d") : secret "infra-operator-webhook-server-cert" not found Dec 02 01:52:57 crc kubenswrapper[4884]: I1202 01:52:57.443539 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd456f66\" (UID: \"2fca9668-22fe-4ea5-a438-f72c6e9c5294\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" Dec 02 01:52:57 crc kubenswrapper[4884]: E1202 01:52:57.443824 4884 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 01:52:57 crc kubenswrapper[4884]: E1202 01:52:57.443938 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert podName:2fca9668-22fe-4ea5-a438-f72c6e9c5294 nodeName:}" failed. No retries permitted until 2025-12-02 01:53:01.443893623 +0000 UTC m=+998.119730507 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd456f66" (UID: "2fca9668-22fe-4ea5-a438-f72c6e9c5294") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 01:52:57 crc kubenswrapper[4884]: I1202 01:52:57.747088 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:52:57 crc kubenswrapper[4884]: I1202 01:52:57.747187 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:52:57 crc kubenswrapper[4884]: E1202 01:52:57.747307 4884 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 01:52:57 crc kubenswrapper[4884]: E1202 01:52:57.747347 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs podName:130cad20-8dfc-4f3d-a00f-d0d5d066ef54 nodeName:}" failed. No retries permitted until 2025-12-02 01:53:01.747337638 +0000 UTC m=+998.423174522 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs") pod "openstack-operator-controller-manager-668dd49799-k2fql" (UID: "130cad20-8dfc-4f3d-a00f-d0d5d066ef54") : secret "webhook-server-cert" not found Dec 02 01:52:57 crc kubenswrapper[4884]: E1202 01:52:57.747632 4884 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 01:52:57 crc kubenswrapper[4884]: E1202 01:52:57.747659 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs podName:130cad20-8dfc-4f3d-a00f-d0d5d066ef54 nodeName:}" failed. No retries permitted until 2025-12-02 01:53:01.747650536 +0000 UTC m=+998.423487430 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs") pod "openstack-operator-controller-manager-668dd49799-k2fql" (UID: "130cad20-8dfc-4f3d-a00f-d0d5d066ef54") : secret "metrics-server-cert" not found Dec 02 01:53:01 crc kubenswrapper[4884]: E1202 01:53:01.096577 4884 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 02 01:53:01 crc kubenswrapper[4884]: E1202 01:53:01.097069 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert podName:6770b6e2-f0dc-4a64-96c7-3d0854fadf3d nodeName:}" failed. No retries permitted until 2025-12-02 01:53:09.097055355 +0000 UTC m=+1005.772892239 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert") pod "infra-operator-controller-manager-57548d458d-nhlqg" (UID: "6770b6e2-f0dc-4a64-96c7-3d0854fadf3d") : secret "infra-operator-webhook-server-cert" not found Dec 02 01:53:01 crc kubenswrapper[4884]: I1202 01:53:01.097362 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert\") pod \"infra-operator-controller-manager-57548d458d-nhlqg\" (UID: \"6770b6e2-f0dc-4a64-96c7-3d0854fadf3d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" Dec 02 01:53:01 crc kubenswrapper[4884]: I1202 01:53:01.502872 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd456f66\" (UID: \"2fca9668-22fe-4ea5-a438-f72c6e9c5294\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" Dec 02 01:53:01 crc kubenswrapper[4884]: E1202 01:53:01.503143 4884 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 01:53:01 crc kubenswrapper[4884]: E1202 01:53:01.503252 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert podName:2fca9668-22fe-4ea5-a438-f72c6e9c5294 nodeName:}" failed. No retries permitted until 2025-12-02 01:53:09.503223088 +0000 UTC m=+1006.179060012 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd456f66" (UID: "2fca9668-22fe-4ea5-a438-f72c6e9c5294") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 02 01:53:01 crc kubenswrapper[4884]: I1202 01:53:01.808360 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:53:01 crc kubenswrapper[4884]: E1202 01:53:01.808509 4884 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 01:53:01 crc kubenswrapper[4884]: E1202 01:53:01.808564 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs podName:130cad20-8dfc-4f3d-a00f-d0d5d066ef54 nodeName:}" failed. No retries permitted until 2025-12-02 01:53:09.808550689 +0000 UTC m=+1006.484387573 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs") pod "openstack-operator-controller-manager-668dd49799-k2fql" (UID: "130cad20-8dfc-4f3d-a00f-d0d5d066ef54") : secret "metrics-server-cert" not found Dec 02 01:53:01 crc kubenswrapper[4884]: E1202 01:53:01.808576 4884 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 01:53:01 crc kubenswrapper[4884]: E1202 01:53:01.808614 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs podName:130cad20-8dfc-4f3d-a00f-d0d5d066ef54 nodeName:}" failed. No retries permitted until 2025-12-02 01:53:09.808603271 +0000 UTC m=+1006.484440155 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs") pod "openstack-operator-controller-manager-668dd49799-k2fql" (UID: "130cad20-8dfc-4f3d-a00f-d0d5d066ef54") : secret "webhook-server-cert" not found Dec 02 01:53:01 crc kubenswrapper[4884]: I1202 01:53:01.808511 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:53:08 crc kubenswrapper[4884]: E1202 01:53:08.035842 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9t7r7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6546668bfd-dm29n_openstack-operators(8b6aa6b7-93b2-4ba1-9a01-10c47d21df36): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 01:53:08 crc kubenswrapper[4884]: E1202 01:53:08.037819 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n" podUID="8b6aa6b7-93b2-4ba1-9a01-10c47d21df36" Dec 02 01:53:08 crc kubenswrapper[4884]: E1202 01:53:08.042373 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rnrsd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-668d9c48b9-qpg2b_openstack-operators(15629082-5a45-439e-b98e-57083adea72c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 02 01:53:08 crc kubenswrapper[4884]: E1202 01:53:08.043564 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b" podUID="15629082-5a45-439e-b98e-57083adea72c" Dec 02 01:53:08 crc kubenswrapper[4884]: I1202 01:53:08.110292 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-j49bt" event={"ID":"4cbce330-7938-453f-989a-c29213cf9118","Type":"ContainerStarted","Data":"592f1ab3bbce489e54e5130ca56543e6a400ca6e2b157c5d27bb7bb94d501947"} Dec 02 01:53:08 crc kubenswrapper[4884]: I1202 01:53:08.120381 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-d2df5" event={"ID":"17378750-b1cf-4515-81f4-638404739f0f","Type":"ContainerStarted","Data":"a6f1df8ea6291eaeb9b625b62d918db4919c680c07de1544f67d2d0c0d060e31"} Dec 02 01:53:08 crc kubenswrapper[4884]: I1202 01:53:08.125065 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kszl7" event={"ID":"f3bc5025-5eac-4f60-a65c-c1abda8e6367","Type":"ContainerStarted","Data":"69ce6aaeea96ac31db64cdcca13772cbe9c21a7d815ef31f869a59beadc2df5e"} Dec 02 01:53:08 crc kubenswrapper[4884]: I1202 01:53:08.126312 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b" event={"ID":"15629082-5a45-439e-b98e-57083adea72c","Type":"ContainerStarted","Data":"5bd4a5ce7033abb33bbdb3c0a0ee04d6bddaf792b4cb20b474556893e6ad1e98"} Dec 02 01:53:08 crc kubenswrapper[4884]: I1202 01:53:08.127016 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b" Dec 02 01:53:08 crc kubenswrapper[4884]: E1202 01:53:08.127827 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b" podUID="15629082-5a45-439e-b98e-57083adea72c" Dec 02 01:53:08 crc kubenswrapper[4884]: I1202 01:53:08.128691 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7tjcj" event={"ID":"6c738783-6d27-48e8-8f8e-a34f0f0ecef8","Type":"ContainerStarted","Data":"7c7a85b2032423989dda342c245422eb4b141066a44c31f9a0bb0564cd95f17a"} Dec 02 01:53:08 crc kubenswrapper[4884]: I1202 01:53:08.132576 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h8944" event={"ID":"ccee2404-f917-4dc3-bd71-be8c8aaf5362","Type":"ContainerStarted","Data":"1b120cb8331e521c7a5f81e4993f0763e3477836b4b95fb612c722f4006ab27b"} Dec 02 01:53:08 crc kubenswrapper[4884]: I1202 01:53:08.135541 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-kf879" event={"ID":"c9178c86-b9ac-4d5b-8145-0cc5f0e79157","Type":"ContainerStarted","Data":"b1ff47689c3bdf45a3effcbf6b88e27857b6ea14c2ca27b597364c8bb28688ea"} Dec 02 01:53:08 crc kubenswrapper[4884]: I1202 01:53:08.137877 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4pvr8" event={"ID":"68a5e45a-0241-4dad-97d5-3855e1128b01","Type":"ContainerStarted","Data":"96f8bec3df72702753ac21ced6648d5fa24810db379a50a4106f33c9f876a74a"} Dec 02 01:53:08 crc kubenswrapper[4884]: I1202 01:53:08.144731 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n" event={"ID":"8b6aa6b7-93b2-4ba1-9a01-10c47d21df36","Type":"ContainerStarted","Data":"c79737fe1e4d6bd3df8c8fd9b1e52d0fb09ee1e2e8e735f3c6c8cc59faa6d88e"} Dec 02 01:53:08 crc kubenswrapper[4884]: I1202 01:53:08.145449 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n" Dec 02 01:53:08 crc kubenswrapper[4884]: E1202 01:53:08.146494 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n" podUID="8b6aa6b7-93b2-4ba1-9a01-10c47d21df36" Dec 02 01:53:08 crc kubenswrapper[4884]: I1202 01:53:08.151830 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-kmj5d" event={"ID":"d7fbdcb3-5250-4b03-b182-47506f9e0c50","Type":"ContainerStarted","Data":"b7bff72d2324da44ef85cb70a941a123a294cbf0742ee8e1bdb92300c7b3c0c5"} Dec 02 01:53:08 crc kubenswrapper[4884]: I1202 01:53:08.170331 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fspk2" event={"ID":"50955276-266a-4e3d-9537-0efccdd1028d","Type":"ContainerStarted","Data":"d2fbbeea1b5cb05e1691f0e3c8137af5859526e91123dd7f6237501e4e793ce2"} Dec 02 01:53:08 crc kubenswrapper[4884]: I1202 01:53:08.180267 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nzwwd" event={"ID":"dd662773-0d4d-4682-9fdb-b79667483d76","Type":"ContainerStarted","Data":"0c4233822afbfcd82c1a47e52e05f2ba38b979258269e277239bcf7ec9c988d5"} Dec 02 01:53:08 crc kubenswrapper[4884]: I1202 01:53:08.181539 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vz95x" event={"ID":"58dc93d7-c6f5-4927-9a80-887c561cbe3b","Type":"ContainerStarted","Data":"e24bc10916896c5ee69d5ed6b782ceba2ceb92771bf79968c53a4c03077b7eb7"} Dec 02 01:53:09 crc kubenswrapper[4884]: I1202 01:53:09.131597 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert\") pod \"infra-operator-controller-manager-57548d458d-nhlqg\" (UID: \"6770b6e2-f0dc-4a64-96c7-3d0854fadf3d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" Dec 02 01:53:09 crc kubenswrapper[4884]: I1202 01:53:09.142288 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6770b6e2-f0dc-4a64-96c7-3d0854fadf3d-cert\") pod \"infra-operator-controller-manager-57548d458d-nhlqg\" (UID: \"6770b6e2-f0dc-4a64-96c7-3d0854fadf3d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" Dec 02 01:53:09 crc kubenswrapper[4884]: E1202 01:53:09.198294 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b" podUID="15629082-5a45-439e-b98e-57083adea72c" Dec 02 01:53:09 crc kubenswrapper[4884]: E1202 01:53:09.198295 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n" podUID="8b6aa6b7-93b2-4ba1-9a01-10c47d21df36" Dec 02 01:53:09 crc kubenswrapper[4884]: I1202 01:53:09.437958 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" Dec 02 01:53:09 crc kubenswrapper[4884]: I1202 01:53:09.540714 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd456f66\" (UID: \"2fca9668-22fe-4ea5-a438-f72c6e9c5294\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" Dec 02 01:53:09 crc kubenswrapper[4884]: I1202 01:53:09.560261 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2fca9668-22fe-4ea5-a438-f72c6e9c5294-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd456f66\" (UID: \"2fca9668-22fe-4ea5-a438-f72c6e9c5294\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" Dec 02 01:53:09 crc kubenswrapper[4884]: I1202 01:53:09.578069 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" Dec 02 01:53:09 crc kubenswrapper[4884]: I1202 01:53:09.844150 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:53:09 crc kubenswrapper[4884]: E1202 01:53:09.844299 4884 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 02 01:53:09 crc kubenswrapper[4884]: E1202 01:53:09.844346 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs podName:130cad20-8dfc-4f3d-a00f-d0d5d066ef54 nodeName:}" failed. No retries permitted until 2025-12-02 01:53:25.844332856 +0000 UTC m=+1022.520169740 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs") pod "openstack-operator-controller-manager-668dd49799-k2fql" (UID: "130cad20-8dfc-4f3d-a00f-d0d5d066ef54") : secret "webhook-server-cert" not found Dec 02 01:53:09 crc kubenswrapper[4884]: I1202 01:53:09.844454 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:53:09 crc kubenswrapper[4884]: E1202 01:53:09.844627 4884 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 02 01:53:09 crc kubenswrapper[4884]: E1202 01:53:09.844657 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs podName:130cad20-8dfc-4f3d-a00f-d0d5d066ef54 nodeName:}" failed. No retries permitted until 2025-12-02 01:53:25.844650484 +0000 UTC m=+1022.520487368 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs") pod "openstack-operator-controller-manager-668dd49799-k2fql" (UID: "130cad20-8dfc-4f3d-a00f-d0d5d066ef54") : secret "metrics-server-cert" not found Dec 02 01:53:10 crc kubenswrapper[4884]: I1202 01:53:10.722373 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66"] Dec 02 01:53:10 crc kubenswrapper[4884]: W1202 01:53:10.849861 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fca9668_22fe_4ea5_a438_f72c6e9c5294.slice/crio-83cf51e0a27f0871c3ce59b878d1bd992307888ad0c42b5824f1032c9e827617 WatchSource:0}: Error finding container 83cf51e0a27f0871c3ce59b878d1bd992307888ad0c42b5824f1032c9e827617: Status 404 returned error can't find the container with id 83cf51e0a27f0871c3ce59b878d1bd992307888ad0c42b5824f1032c9e827617 Dec 02 01:53:11 crc kubenswrapper[4884]: I1202 01:53:11.023463 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg"] Dec 02 01:53:11 crc kubenswrapper[4884]: I1202 01:53:11.216154 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" event={"ID":"2fca9668-22fe-4ea5-a438-f72c6e9c5294","Type":"ContainerStarted","Data":"83cf51e0a27f0871c3ce59b878d1bd992307888ad0c42b5824f1032c9e827617"} Dec 02 01:53:11 crc kubenswrapper[4884]: W1202 01:53:11.574849 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6770b6e2_f0dc_4a64_96c7_3d0854fadf3d.slice/crio-88aa8472452b16d70d34ced5be440141d37512e0820a4c842a4a6b490024652c WatchSource:0}: Error finding container 88aa8472452b16d70d34ced5be440141d37512e0820a4c842a4a6b490024652c: Status 404 returned error can't find the container with id 88aa8472452b16d70d34ced5be440141d37512e0820a4c842a4a6b490024652c Dec 02 01:53:12 crc kubenswrapper[4884]: I1202 01:53:12.229413 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" event={"ID":"6770b6e2-f0dc-4a64-96c7-3d0854fadf3d","Type":"ContainerStarted","Data":"88aa8472452b16d70d34ced5be440141d37512e0820a4c842a4a6b490024652c"} Dec 02 01:53:13 crc kubenswrapper[4884]: I1202 01:53:13.430473 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b" Dec 02 01:53:13 crc kubenswrapper[4884]: E1202 01:53:13.432814 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b" podUID="15629082-5a45-439e-b98e-57083adea72c" Dec 02 01:53:13 crc kubenswrapper[4884]: I1202 01:53:13.834408 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n" Dec 02 01:53:13 crc kubenswrapper[4884]: E1202 01:53:13.841150 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n" podUID="8b6aa6b7-93b2-4ba1-9a01-10c47d21df36" Dec 02 01:53:16 crc kubenswrapper[4884]: I1202 01:53:16.971499 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:53:16 crc kubenswrapper[4884]: I1202 01:53:16.971884 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:53:25 crc kubenswrapper[4884]: I1202 01:53:25.924653 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:53:25 crc kubenswrapper[4884]: I1202 01:53:25.926055 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:53:25 crc kubenswrapper[4884]: I1202 01:53:25.932939 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-metrics-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:53:25 crc kubenswrapper[4884]: I1202 01:53:25.939930 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/130cad20-8dfc-4f3d-a00f-d0d5d066ef54-webhook-certs\") pod \"openstack-operator-controller-manager-668dd49799-k2fql\" (UID: \"130cad20-8dfc-4f3d-a00f-d0d5d066ef54\") " pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:53:26 crc kubenswrapper[4884]: I1202 01:53:26.121612 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-stmn2" Dec 02 01:53:26 crc kubenswrapper[4884]: I1202 01:53:26.129814 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:53:28 crc kubenswrapper[4884]: E1202 01:53:28.700978 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385" Dec 02 01:53:28 crc kubenswrapper[4884]: E1202 01:53:28.701575 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4j5qj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-x5hdr_openstack-operators(c7c9137a-0bc9-41e0-9dc0-f5976861a3b2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 01:53:30 crc kubenswrapper[4884]: E1202 01:53:30.441395 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 02 01:53:30 crc kubenswrapper[4884]: E1202 01:53:30.443110 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jq9ks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-qjjv8_openstack-operators(d7c0cc8c-1763-443f-a2a0-5d76d093b019): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 01:53:30 crc kubenswrapper[4884]: E1202 01:53:30.444516 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8" podUID="d7c0cc8c-1763-443f-a2a0-5d76d093b019" Dec 02 01:53:30 crc kubenswrapper[4884]: I1202 01:53:30.720973 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql"] Dec 02 01:53:30 crc kubenswrapper[4884]: W1202 01:53:30.776204 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod130cad20_8dfc_4f3d_a00f_d0d5d066ef54.slice/crio-ad162127e43a32ec2fcb52e048885808f702a6fff81fb72519a47f0705c6e14d WatchSource:0}: Error finding container ad162127e43a32ec2fcb52e048885808f702a6fff81fb72519a47f0705c6e14d: Status 404 returned error can't find the container with id ad162127e43a32ec2fcb52e048885808f702a6fff81fb72519a47f0705c6e14d Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.393634 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf" event={"ID":"8e4db64b-9154-48cd-bd3c-463038b4dd93","Type":"ContainerStarted","Data":"9e52ebaea72e22ad66c13edbff8bbf0ef6aad2f1ef590ef30213c131e92ddc14"} Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.416203 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-j49bt" event={"ID":"4cbce330-7938-453f-989a-c29213cf9118","Type":"ContainerStarted","Data":"be99bb45ea47a9395be798ea056503df765013168c69ea5a1d1261072b27cb0c"} Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.416631 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-j49bt" Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.418561 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-j49bt" Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.425709 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7tjcj" event={"ID":"6c738783-6d27-48e8-8f8e-a34f0f0ecef8","Type":"ContainerStarted","Data":"58e7cb5770e3e0e9f969127213440d508cb4069000f867642122f471383e0b8a"} Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.426539 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7tjcj" Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.430646 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7tjcj" Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.436062 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw" event={"ID":"4446fb10-1203-4063-affd-bc81cbfa4238","Type":"ContainerStarted","Data":"15e7c811afba0a38ac2f831c675a25e63ca7b7606777aa2589705eb0e1607a1a"} Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.446924 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-j49bt" podStartSLOduration=2.9357066339999998 podStartE2EDuration="38.446906921s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:54.891023093 +0000 UTC m=+991.566859977" lastFinishedPulling="2025-12-02 01:53:30.40222339 +0000 UTC m=+1027.078060264" observedRunningTime="2025-12-02 01:53:31.436586263 +0000 UTC m=+1028.112423147" watchObservedRunningTime="2025-12-02 01:53:31.446906921 +0000 UTC m=+1028.122743805" Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.473918 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" event={"ID":"2fca9668-22fe-4ea5-a438-f72c6e9c5294","Type":"ContainerStarted","Data":"a293808d6a034295110ddd4d66994b5debe8c2c26621f18381af934256132ec3"} Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.509901 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n" event={"ID":"8b6aa6b7-93b2-4ba1-9a01-10c47d21df36","Type":"ContainerStarted","Data":"0ea8b9d8d6bb65cd7d1b3199a43f778319d18804c2430bf0e4aac374dd814a05"} Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.517361 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-7tjcj" podStartSLOduration=4.660377903 podStartE2EDuration="38.517337842s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:54.804920681 +0000 UTC m=+991.480757565" lastFinishedPulling="2025-12-02 01:53:28.66188061 +0000 UTC m=+1025.337717504" observedRunningTime="2025-12-02 01:53:31.475798754 +0000 UTC m=+1028.151635638" watchObservedRunningTime="2025-12-02 01:53:31.517337842 +0000 UTC m=+1028.193174726" Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.521567 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" event={"ID":"130cad20-8dfc-4f3d-a00f-d0d5d066ef54","Type":"ContainerStarted","Data":"ad162127e43a32ec2fcb52e048885808f702a6fff81fb72519a47f0705c6e14d"} Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.525599 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww" event={"ID":"edf5e61d-2bf9-4eb6-9363-88e76f4e9506","Type":"ContainerStarted","Data":"01a92c88ed1181ce9806c26b1a57ae5ed8eb3bb73931e101022658837f409b11"} Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.527732 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-kf879" event={"ID":"c9178c86-b9ac-4d5b-8145-0cc5f0e79157","Type":"ContainerStarted","Data":"080cf8c3b4778f50fde84c0094d6dbc7f966f561af3b882a7cb263ac224ede8d"} Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.528344 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-kf879" Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.530084 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-kf879" Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.537853 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n" event={"ID":"264f2ec8-0c0a-4fce-8ed3-eb8fafae0621","Type":"ContainerStarted","Data":"6b0ab0d96b42d3a3c0e31f8bd80c0137edd288d2c4a801d70636971f917c777f"} Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.557758 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6546668bfd-dm29n" podStartSLOduration=26.331607575 podStartE2EDuration="38.557728561s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:55.203111854 +0000 UTC m=+991.878948738" lastFinishedPulling="2025-12-02 01:53:07.42923284 +0000 UTC m=+1004.105069724" observedRunningTime="2025-12-02 01:53:31.537141647 +0000 UTC m=+1028.212978531" watchObservedRunningTime="2025-12-02 01:53:31.557728561 +0000 UTC m=+1028.233565445" Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.581027 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-kmj5d" event={"ID":"d7fbdcb3-5250-4b03-b182-47506f9e0c50","Type":"ContainerStarted","Data":"4f2b5a135e9f092a718c5c07d2a7afe6803e5ea7bd6c61099d34cd1ce3c7fe1e"} Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.582306 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-kmj5d" Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.583319 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-kf879" podStartSLOduration=15.763413209 podStartE2EDuration="38.58329455s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:55.219540405 +0000 UTC m=+991.895377289" lastFinishedPulling="2025-12-02 01:53:18.039421746 +0000 UTC m=+1014.715258630" observedRunningTime="2025-12-02 01:53:31.560790078 +0000 UTC m=+1028.236626962" watchObservedRunningTime="2025-12-02 01:53:31.58329455 +0000 UTC m=+1028.259131444" Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.589698 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-kmj5d" Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.615311 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-546d4bdf48-kmj5d" podStartSLOduration=15.095503434 podStartE2EDuration="38.6152966s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:55.130603382 +0000 UTC m=+991.806440276" lastFinishedPulling="2025-12-02 01:53:18.650396568 +0000 UTC m=+1015.326233442" observedRunningTime="2025-12-02 01:53:31.611809703 +0000 UTC m=+1028.287646597" watchObservedRunningTime="2025-12-02 01:53:31.6152966 +0000 UTC m=+1028.291133484" Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.679984 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fspk2" Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.680530 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fspk2" Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.680610 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fspk2" event={"ID":"50955276-266a-4e3d-9537-0efccdd1028d","Type":"ContainerStarted","Data":"fe916efe98978109e6cf296ba9ab6f7e3badd9bd2ec945f6bafc7f09b7e4bab1"} Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.694924 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" event={"ID":"6770b6e2-f0dc-4a64-96c7-3d0854fadf3d","Type":"ContainerStarted","Data":"5be788d097304f38506d3a4b7797d67fafd5722ef61eb45e6620b5edb72ef58d"} Dec 02 01:53:31 crc kubenswrapper[4884]: I1202 01:53:31.748692 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-fspk2" podStartSLOduration=3.870230388 podStartE2EDuration="39.748674704s" podCreationTimestamp="2025-12-02 01:52:52 +0000 UTC" firstStartedPulling="2025-12-02 01:52:54.525549798 +0000 UTC m=+991.201386682" lastFinishedPulling="2025-12-02 01:53:30.403994114 +0000 UTC m=+1027.079830998" observedRunningTime="2025-12-02 01:53:31.714759517 +0000 UTC m=+1028.390596401" watchObservedRunningTime="2025-12-02 01:53:31.748674704 +0000 UTC m=+1028.424511588" Dec 02 01:53:31 crc kubenswrapper[4884]: E1202 01:53:31.935250 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" podUID="c7c9137a-0bc9-41e0-9dc0-f5976861a3b2" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.701413 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-d2df5" event={"ID":"17378750-b1cf-4515-81f4-638404739f0f","Type":"ContainerStarted","Data":"89af0b2b64dba58eb68dec8f72fd767109883c92c581a33c2de09fd87c910aae"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.702545 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-d2df5" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.703571 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n" event={"ID":"264f2ec8-0c0a-4fce-8ed3-eb8fafae0621","Type":"ContainerStarted","Data":"05b358ab820c9d0c8e79a459194a945a375a9167da2959967806bd3bfcaec58a"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.703974 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.705512 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8" event={"ID":"3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f","Type":"ContainerStarted","Data":"f5944ddf9044504160ba25e0282427058ab07b25be36fe12e1b18d64f4f60cb8"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.705537 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8" event={"ID":"3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f","Type":"ContainerStarted","Data":"a4894f5e5967eb25efb4bf6d24009461bc6fbfc4d28f9c274a1309157d3a3d27"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.705885 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.706450 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-d2df5" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.707571 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" event={"ID":"2fca9668-22fe-4ea5-a438-f72c6e9c5294","Type":"ContainerStarted","Data":"c60a5b1fd15ea6749333837bddcaf08ca96456084bbb882e74cba13fe0a28b03"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.707951 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.709364 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" event={"ID":"c7c9137a-0bc9-41e0-9dc0-f5976861a3b2","Type":"ContainerStarted","Data":"ab686f29157a859b81975f7777033f0a0c07881b37c3d8c39e876eb472905157"} Dec 02 01:53:32 crc kubenswrapper[4884]: E1202 01:53:32.710353 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" podUID="c7c9137a-0bc9-41e0-9dc0-f5976861a3b2" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.711428 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" event={"ID":"130cad20-8dfc-4f3d-a00f-d0d5d066ef54","Type":"ContainerStarted","Data":"4a8258a8736b0b1272ec4ad356399b690c91752db2ef357364c2586aa2096820"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.711565 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.713190 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf" event={"ID":"8e4db64b-9154-48cd-bd3c-463038b4dd93","Type":"ContainerStarted","Data":"28f210c04c2e422e13bb4813a9619e8adcbdaa9e60c653358564c67e85192e4b"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.713303 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.714628 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vz95x" event={"ID":"58dc93d7-c6f5-4927-9a80-887c561cbe3b","Type":"ContainerStarted","Data":"316e9feca3e7f056b1dc25cd4d86f5d1eea794496b409bce45ebfb3dbfc5459b"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.714727 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vz95x" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.716129 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vz95x" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.716334 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww" event={"ID":"edf5e61d-2bf9-4eb6-9363-88e76f4e9506","Type":"ContainerStarted","Data":"a96fa744d1421721afc0ebf6ded6228d10f92c8b4b08e02e293d7c9f03376383"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.716419 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.718063 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h8944" event={"ID":"ccee2404-f917-4dc3-bd71-be8c8aaf5362","Type":"ContainerStarted","Data":"ab2ab1e5d936a46458e5e7f02723122a598a15a0f4c667630c995127f7b67ba3"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.719568 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h8944" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.721172 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h8944" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.721323 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4pvr8" event={"ID":"68a5e45a-0241-4dad-97d5-3855e1128b01","Type":"ContainerStarted","Data":"9cfb8a585de5bd8f2b1878e1de45edbb3742744cee0857ed0be9cca0d555746f"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.722556 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4pvr8" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.723269 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4pvr8" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.723949 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw" event={"ID":"4446fb10-1203-4063-affd-bc81cbfa4238","Type":"ContainerStarted","Data":"fec983820e7b0370fa7f89abb15753293ccaee9261465c5f8efaf38f0b7c451a"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.724076 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.727142 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b" event={"ID":"15629082-5a45-439e-b98e-57083adea72c","Type":"ContainerStarted","Data":"c7b796d61ae5c767a0671a15f8d7af95af2819d8a663fdc2ce889bd23fae7880"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.730974 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-d2df5" podStartSLOduration=4.433346359 podStartE2EDuration="39.730959817s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:55.114403057 +0000 UTC m=+991.790239941" lastFinishedPulling="2025-12-02 01:53:30.412016515 +0000 UTC m=+1027.087853399" observedRunningTime="2025-12-02 01:53:32.726415023 +0000 UTC m=+1029.402251897" watchObservedRunningTime="2025-12-02 01:53:32.730959817 +0000 UTC m=+1029.406796701" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.732940 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nzwwd" event={"ID":"dd662773-0d4d-4682-9fdb-b79667483d76","Type":"ContainerStarted","Data":"f20c47bc313482ad64e5de8c1a7d72bf4e310e4f6c2061309eef64f7c58d6310"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.734772 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nzwwd" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.736026 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nzwwd" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.736568 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kszl7" event={"ID":"f3bc5025-5eac-4f60-a65c-c1abda8e6367","Type":"ContainerStarted","Data":"582b885a529d3733c187de2375c02ff36b38f033a35d68f7d8caec1a19e139ee"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.736772 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kszl7" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.738254 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kszl7" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.739295 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" event={"ID":"6770b6e2-f0dc-4a64-96c7-3d0854fadf3d","Type":"ContainerStarted","Data":"d6d98f7d2acb5472ac2ed8373c8f0a955afe5d3733a3dddb984fd84c6b0cac23"} Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.739319 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.747075 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw" podStartSLOduration=17.508225492 podStartE2EDuration="39.74706473s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:55.248078368 +0000 UTC m=+991.923915252" lastFinishedPulling="2025-12-02 01:53:17.486917596 +0000 UTC m=+1014.162754490" observedRunningTime="2025-12-02 01:53:32.744805963 +0000 UTC m=+1029.420642857" watchObservedRunningTime="2025-12-02 01:53:32.74706473 +0000 UTC m=+1029.422901614" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.773229 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf" podStartSLOduration=4.619314987 podStartE2EDuration="39.773217003s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:55.24777721 +0000 UTC m=+991.923614094" lastFinishedPulling="2025-12-02 01:53:30.401679206 +0000 UTC m=+1027.077516110" observedRunningTime="2025-12-02 01:53:32.76908149 +0000 UTC m=+1029.444918374" watchObservedRunningTime="2025-12-02 01:53:32.773217003 +0000 UTC m=+1029.449053887" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.816997 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-vz95x" podStartSLOduration=4.618770894 podStartE2EDuration="39.816982697s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:55.203459603 +0000 UTC m=+991.879296487" lastFinishedPulling="2025-12-02 01:53:30.401671406 +0000 UTC m=+1027.077508290" observedRunningTime="2025-12-02 01:53:32.811969762 +0000 UTC m=+1029.487806656" watchObservedRunningTime="2025-12-02 01:53:32.816982697 +0000 UTC m=+1029.492819571" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.866709 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-4pvr8" podStartSLOduration=4.649718797 podStartE2EDuration="39.866681779s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:55.20373734 +0000 UTC m=+991.879574224" lastFinishedPulling="2025-12-02 01:53:30.420700322 +0000 UTC m=+1027.096537206" observedRunningTime="2025-12-02 01:53:32.866665479 +0000 UTC m=+1029.542502373" watchObservedRunningTime="2025-12-02 01:53:32.866681779 +0000 UTC m=+1029.542518693" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.925541 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-h8944" podStartSLOduration=4.234912078 podStartE2EDuration="39.92552811s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:54.835949546 +0000 UTC m=+991.511786430" lastFinishedPulling="2025-12-02 01:53:30.526565578 +0000 UTC m=+1027.202402462" observedRunningTime="2025-12-02 01:53:32.92390435 +0000 UTC m=+1029.599741234" watchObservedRunningTime="2025-12-02 01:53:32.92552811 +0000 UTC m=+1029.601364994" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.955032 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8" podStartSLOduration=4.760208589 podStartE2EDuration="39.955018477s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:55.2245639 +0000 UTC m=+991.900400784" lastFinishedPulling="2025-12-02 01:53:30.419373768 +0000 UTC m=+1027.095210672" observedRunningTime="2025-12-02 01:53:32.950850383 +0000 UTC m=+1029.626687277" watchObservedRunningTime="2025-12-02 01:53:32.955018477 +0000 UTC m=+1029.630855361" Dec 02 01:53:32 crc kubenswrapper[4884]: I1202 01:53:32.975392 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n" podStartSLOduration=17.735556653 podStartE2EDuration="39.975370416s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:55.246243122 +0000 UTC m=+991.922080006" lastFinishedPulling="2025-12-02 01:53:17.486056845 +0000 UTC m=+1014.161893769" observedRunningTime="2025-12-02 01:53:32.974523905 +0000 UTC m=+1029.650360789" watchObservedRunningTime="2025-12-02 01:53:32.975370416 +0000 UTC m=+1029.651207320" Dec 02 01:53:33 crc kubenswrapper[4884]: I1202 01:53:33.017461 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" podStartSLOduration=20.452256608 podStartE2EDuration="40.017445228s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:53:10.855704946 +0000 UTC m=+1007.531541830" lastFinishedPulling="2025-12-02 01:53:30.420893556 +0000 UTC m=+1027.096730450" observedRunningTime="2025-12-02 01:53:33.012042513 +0000 UTC m=+1029.687879407" watchObservedRunningTime="2025-12-02 01:53:33.017445228 +0000 UTC m=+1029.693282112" Dec 02 01:53:33 crc kubenswrapper[4884]: I1202 01:53:33.036435 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww" podStartSLOduration=17.166095229 podStartE2EDuration="40.036416792s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:55.22015944 +0000 UTC m=+991.895996324" lastFinishedPulling="2025-12-02 01:53:18.090481003 +0000 UTC m=+1014.766317887" observedRunningTime="2025-12-02 01:53:33.035094669 +0000 UTC m=+1029.710931553" watchObservedRunningTime="2025-12-02 01:53:33.036416792 +0000 UTC m=+1029.712253676" Dec 02 01:53:33 crc kubenswrapper[4884]: I1202 01:53:33.090228 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" podStartSLOduration=40.090213107 podStartE2EDuration="40.090213107s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:53:33.089879788 +0000 UTC m=+1029.765716682" watchObservedRunningTime="2025-12-02 01:53:33.090213107 +0000 UTC m=+1029.766049991" Dec 02 01:53:33 crc kubenswrapper[4884]: I1202 01:53:33.114840 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-kszl7" podStartSLOduration=4.832225319 podStartE2EDuration="40.114822692s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:55.139429732 +0000 UTC m=+991.815266626" lastFinishedPulling="2025-12-02 01:53:30.422027115 +0000 UTC m=+1027.097863999" observedRunningTime="2025-12-02 01:53:33.114469903 +0000 UTC m=+1029.790306807" watchObservedRunningTime="2025-12-02 01:53:33.114822692 +0000 UTC m=+1029.790659576" Dec 02 01:53:33 crc kubenswrapper[4884]: I1202 01:53:33.134317 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-668d9c48b9-qpg2b" podStartSLOduration=27.590593994 podStartE2EDuration="40.134298809s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:54.836303975 +0000 UTC m=+991.512140859" lastFinishedPulling="2025-12-02 01:53:07.38000879 +0000 UTC m=+1004.055845674" observedRunningTime="2025-12-02 01:53:33.132823312 +0000 UTC m=+1029.808660206" watchObservedRunningTime="2025-12-02 01:53:33.134298809 +0000 UTC m=+1029.810135693" Dec 02 01:53:33 crc kubenswrapper[4884]: I1202 01:53:33.155168 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-nzwwd" podStartSLOduration=4.973185932 podStartE2EDuration="40.15515393s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:55.219964795 +0000 UTC m=+991.895801679" lastFinishedPulling="2025-12-02 01:53:30.401932793 +0000 UTC m=+1027.077769677" observedRunningTime="2025-12-02 01:53:33.154672688 +0000 UTC m=+1029.830509572" watchObservedRunningTime="2025-12-02 01:53:33.15515393 +0000 UTC m=+1029.830990814" Dec 02 01:53:33 crc kubenswrapper[4884]: I1202 01:53:33.179175 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" podStartSLOduration=21.338060459 podStartE2EDuration="40.17916016s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:53:11.577844497 +0000 UTC m=+1008.253681381" lastFinishedPulling="2025-12-02 01:53:30.418944198 +0000 UTC m=+1027.094781082" observedRunningTime="2025-12-02 01:53:33.177811096 +0000 UTC m=+1029.853647990" watchObservedRunningTime="2025-12-02 01:53:33.17916016 +0000 UTC m=+1029.854997044" Dec 02 01:53:36 crc kubenswrapper[4884]: I1202 01:53:36.140488 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-668dd49799-k2fql" Dec 02 01:53:39 crc kubenswrapper[4884]: I1202 01:53:39.448289 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-nhlqg" Dec 02 01:53:39 crc kubenswrapper[4884]: I1202 01:53:39.584925 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd456f66" Dec 02 01:53:43 crc kubenswrapper[4884]: I1202 01:53:43.896440 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-r8hgf" Dec 02 01:53:43 crc kubenswrapper[4884]: I1202 01:53:43.922581 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-w5xb8" Dec 02 01:53:44 crc kubenswrapper[4884]: I1202 01:53:44.068761 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-znmww" Dec 02 01:53:44 crc kubenswrapper[4884]: I1202 01:53:44.283854 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-nnfmw" Dec 02 01:53:44 crc kubenswrapper[4884]: I1202 01:53:44.309734 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-84f44f5d76-fz96n" Dec 02 01:53:45 crc kubenswrapper[4884]: E1202 01:53:45.649157 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8" podUID="d7c0cc8c-1763-443f-a2a0-5d76d093b019" Dec 02 01:53:46 crc kubenswrapper[4884]: E1202 01:53:46.615989 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" podUID="c7c9137a-0bc9-41e0-9dc0-f5976861a3b2" Dec 02 01:53:46 crc kubenswrapper[4884]: I1202 01:53:46.971973 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:53:46 crc kubenswrapper[4884]: I1202 01:53:46.972067 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:53:46 crc kubenswrapper[4884]: I1202 01:53:46.972143 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:53:46 crc kubenswrapper[4884]: I1202 01:53:46.973164 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7b403650e63fc1ca4927612983b11a4a56c5610f07757026da030e5a5f4be8dc"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 01:53:46 crc kubenswrapper[4884]: I1202 01:53:46.973295 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://7b403650e63fc1ca4927612983b11a4a56c5610f07757026da030e5a5f4be8dc" gracePeriod=600 Dec 02 01:53:47 crc kubenswrapper[4884]: I1202 01:53:47.999192 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="7b403650e63fc1ca4927612983b11a4a56c5610f07757026da030e5a5f4be8dc" exitCode=0 Dec 02 01:53:47 crc kubenswrapper[4884]: I1202 01:53:47.999287 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"7b403650e63fc1ca4927612983b11a4a56c5610f07757026da030e5a5f4be8dc"} Dec 02 01:53:48 crc kubenswrapper[4884]: I1202 01:53:47.999574 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"58124d3cd233d6efb787d810e5402863819796c49f3516cfa49fd6248e566581"} Dec 02 01:53:48 crc kubenswrapper[4884]: I1202 01:53:47.999600 4884 scope.go:117] "RemoveContainer" containerID="ec6281f97479113b2e5dc03c077596d5c0e07b1ee23616b661b78bdd0a2d70c8" Dec 02 01:53:58 crc kubenswrapper[4884]: I1202 01:53:58.100450 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8" event={"ID":"d7c0cc8c-1763-443f-a2a0-5d76d093b019","Type":"ContainerStarted","Data":"d8e07a44b98f8f0185252b05bece3f749ba4a898bea1ceb112f34352a0793a7d"} Dec 02 01:53:58 crc kubenswrapper[4884]: I1202 01:53:58.147172 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-qjjv8" podStartSLOduration=2.997625719 podStartE2EDuration="1m5.147110012s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:55.252399476 +0000 UTC m=+991.928236360" lastFinishedPulling="2025-12-02 01:53:57.401883729 +0000 UTC m=+1054.077720653" observedRunningTime="2025-12-02 01:53:58.139436988 +0000 UTC m=+1054.815273932" watchObservedRunningTime="2025-12-02 01:53:58.147110012 +0000 UTC m=+1054.822946926" Dec 02 01:54:01 crc kubenswrapper[4884]: I1202 01:54:01.126937 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" event={"ID":"c7c9137a-0bc9-41e0-9dc0-f5976861a3b2","Type":"ContainerStarted","Data":"e3c6c937ba8cbafa629c46c8b7e869b3e2bd82221d16b7ac2d8dd7419c9d235b"} Dec 02 01:54:01 crc kubenswrapper[4884]: I1202 01:54:01.128043 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" Dec 02 01:54:01 crc kubenswrapper[4884]: I1202 01:54:01.164856 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" podStartSLOduration=3.165746153 podStartE2EDuration="1m8.164830546s" podCreationTimestamp="2025-12-02 01:52:53 +0000 UTC" firstStartedPulling="2025-12-02 01:52:55.252384106 +0000 UTC m=+991.928220990" lastFinishedPulling="2025-12-02 01:54:00.251468499 +0000 UTC m=+1056.927305383" observedRunningTime="2025-12-02 01:54:01.153377146 +0000 UTC m=+1057.829214030" watchObservedRunningTime="2025-12-02 01:54:01.164830546 +0000 UTC m=+1057.840667470" Dec 02 01:54:14 crc kubenswrapper[4884]: I1202 01:54:14.144012 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-x5hdr" Dec 02 01:54:31 crc kubenswrapper[4884]: I1202 01:54:31.896347 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6zx6q"] Dec 02 01:54:31 crc kubenswrapper[4884]: I1202 01:54:31.897954 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-6zx6q" Dec 02 01:54:31 crc kubenswrapper[4884]: I1202 01:54:31.903236 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 02 01:54:31 crc kubenswrapper[4884]: I1202 01:54:31.904031 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 02 01:54:31 crc kubenswrapper[4884]: I1202 01:54:31.904203 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 02 01:54:31 crc kubenswrapper[4884]: I1202 01:54:31.904354 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-2tdtl" Dec 02 01:54:31 crc kubenswrapper[4884]: I1202 01:54:31.940307 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6zx6q"] Dec 02 01:54:31 crc kubenswrapper[4884]: I1202 01:54:31.971013 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wf8w2"] Dec 02 01:54:31 crc kubenswrapper[4884]: I1202 01:54:31.972882 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" Dec 02 01:54:31 crc kubenswrapper[4884]: I1202 01:54:31.975530 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 02 01:54:31 crc kubenswrapper[4884]: I1202 01:54:31.990059 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wf8w2"] Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.044675 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c1ac317-3e2a-435c-bd07-933cb4e4fcc3-config\") pod \"dnsmasq-dns-675f4bcbfc-6zx6q\" (UID: \"4c1ac317-3e2a-435c-bd07-933cb4e4fcc3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6zx6q" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.044818 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk7kx\" (UniqueName: \"kubernetes.io/projected/4c1ac317-3e2a-435c-bd07-933cb4e4fcc3-kube-api-access-tk7kx\") pod \"dnsmasq-dns-675f4bcbfc-6zx6q\" (UID: \"4c1ac317-3e2a-435c-bd07-933cb4e4fcc3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6zx6q" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.145490 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c1ac317-3e2a-435c-bd07-933cb4e4fcc3-config\") pod \"dnsmasq-dns-675f4bcbfc-6zx6q\" (UID: \"4c1ac317-3e2a-435c-bd07-933cb4e4fcc3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6zx6q" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.146442 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj9fh\" (UniqueName: \"kubernetes.io/projected/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-kube-api-access-cj9fh\") pod \"dnsmasq-dns-78dd6ddcc-wf8w2\" (UID: \"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.146387 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c1ac317-3e2a-435c-bd07-933cb4e4fcc3-config\") pod \"dnsmasq-dns-675f4bcbfc-6zx6q\" (UID: \"4c1ac317-3e2a-435c-bd07-933cb4e4fcc3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6zx6q" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.146592 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-wf8w2\" (UID: \"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.146642 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-config\") pod \"dnsmasq-dns-78dd6ddcc-wf8w2\" (UID: \"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.146684 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk7kx\" (UniqueName: \"kubernetes.io/projected/4c1ac317-3e2a-435c-bd07-933cb4e4fcc3-kube-api-access-tk7kx\") pod \"dnsmasq-dns-675f4bcbfc-6zx6q\" (UID: \"4c1ac317-3e2a-435c-bd07-933cb4e4fcc3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6zx6q" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.167832 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk7kx\" (UniqueName: \"kubernetes.io/projected/4c1ac317-3e2a-435c-bd07-933cb4e4fcc3-kube-api-access-tk7kx\") pod \"dnsmasq-dns-675f4bcbfc-6zx6q\" (UID: \"4c1ac317-3e2a-435c-bd07-933cb4e4fcc3\") " pod="openstack/dnsmasq-dns-675f4bcbfc-6zx6q" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.223207 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-6zx6q" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.247728 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj9fh\" (UniqueName: \"kubernetes.io/projected/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-kube-api-access-cj9fh\") pod \"dnsmasq-dns-78dd6ddcc-wf8w2\" (UID: \"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.247882 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-wf8w2\" (UID: \"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.247933 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-config\") pod \"dnsmasq-dns-78dd6ddcc-wf8w2\" (UID: \"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.249096 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-wf8w2\" (UID: \"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.251534 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-config\") pod \"dnsmasq-dns-78dd6ddcc-wf8w2\" (UID: \"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.264241 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj9fh\" (UniqueName: \"kubernetes.io/projected/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-kube-api-access-cj9fh\") pod \"dnsmasq-dns-78dd6ddcc-wf8w2\" (UID: \"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.303105 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.525670 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wf8w2"] Dec 02 01:54:32 crc kubenswrapper[4884]: W1202 01:54:32.666569 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c1ac317_3e2a_435c_bd07_933cb4e4fcc3.slice/crio-826112fbf353ff452a4e445197d3ee5a6222fdfd6aa4e4d10827f7f0e40761da WatchSource:0}: Error finding container 826112fbf353ff452a4e445197d3ee5a6222fdfd6aa4e4d10827f7f0e40761da: Status 404 returned error can't find the container with id 826112fbf353ff452a4e445197d3ee5a6222fdfd6aa4e4d10827f7f0e40761da Dec 02 01:54:32 crc kubenswrapper[4884]: I1202 01:54:32.669263 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6zx6q"] Dec 02 01:54:33 crc kubenswrapper[4884]: I1202 01:54:33.454671 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" event={"ID":"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968","Type":"ContainerStarted","Data":"eaeb9f8e69c5a405a8e5e7b7f1f54d47684849a126e6158b9d0aa425580d866c"} Dec 02 01:54:33 crc kubenswrapper[4884]: I1202 01:54:33.457158 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-6zx6q" event={"ID":"4c1ac317-3e2a-435c-bd07-933cb4e4fcc3","Type":"ContainerStarted","Data":"826112fbf353ff452a4e445197d3ee5a6222fdfd6aa4e4d10827f7f0e40761da"} Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.658659 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6zx6q"] Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.679917 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qcp96"] Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.681075 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qcp96" Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.692496 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qcp96"] Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.698785 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df8fd\" (UniqueName: \"kubernetes.io/projected/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-kube-api-access-df8fd\") pod \"dnsmasq-dns-666b6646f7-qcp96\" (UID: \"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214\") " pod="openstack/dnsmasq-dns-666b6646f7-qcp96" Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.698835 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-config\") pod \"dnsmasq-dns-666b6646f7-qcp96\" (UID: \"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214\") " pod="openstack/dnsmasq-dns-666b6646f7-qcp96" Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.698856 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-dns-svc\") pod \"dnsmasq-dns-666b6646f7-qcp96\" (UID: \"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214\") " pod="openstack/dnsmasq-dns-666b6646f7-qcp96" Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.802117 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-config\") pod \"dnsmasq-dns-666b6646f7-qcp96\" (UID: \"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214\") " pod="openstack/dnsmasq-dns-666b6646f7-qcp96" Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.802166 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-dns-svc\") pod \"dnsmasq-dns-666b6646f7-qcp96\" (UID: \"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214\") " pod="openstack/dnsmasq-dns-666b6646f7-qcp96" Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.802259 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df8fd\" (UniqueName: \"kubernetes.io/projected/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-kube-api-access-df8fd\") pod \"dnsmasq-dns-666b6646f7-qcp96\" (UID: \"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214\") " pod="openstack/dnsmasq-dns-666b6646f7-qcp96" Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.803330 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-config\") pod \"dnsmasq-dns-666b6646f7-qcp96\" (UID: \"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214\") " pod="openstack/dnsmasq-dns-666b6646f7-qcp96" Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.803844 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-dns-svc\") pod \"dnsmasq-dns-666b6646f7-qcp96\" (UID: \"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214\") " pod="openstack/dnsmasq-dns-666b6646f7-qcp96" Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.821572 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df8fd\" (UniqueName: \"kubernetes.io/projected/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-kube-api-access-df8fd\") pod \"dnsmasq-dns-666b6646f7-qcp96\" (UID: \"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214\") " pod="openstack/dnsmasq-dns-666b6646f7-qcp96" Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.958716 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wf8w2"] Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.985833 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tf8jx"] Dec 02 01:54:34 crc kubenswrapper[4884]: I1202 01:54:34.986935 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.003189 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qcp96" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.005255 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tf8jx"] Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.009077 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd9xn\" (UniqueName: \"kubernetes.io/projected/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-kube-api-access-bd9xn\") pod \"dnsmasq-dns-57d769cc4f-tf8jx\" (UID: \"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d\") " pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.009125 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-tf8jx\" (UID: \"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d\") " pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.009150 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-config\") pod \"dnsmasq-dns-57d769cc4f-tf8jx\" (UID: \"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d\") " pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.112347 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd9xn\" (UniqueName: \"kubernetes.io/projected/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-kube-api-access-bd9xn\") pod \"dnsmasq-dns-57d769cc4f-tf8jx\" (UID: \"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d\") " pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.112402 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-tf8jx\" (UID: \"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d\") " pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.112428 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-config\") pod \"dnsmasq-dns-57d769cc4f-tf8jx\" (UID: \"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d\") " pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.113590 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-config\") pod \"dnsmasq-dns-57d769cc4f-tf8jx\" (UID: \"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d\") " pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.114326 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-tf8jx\" (UID: \"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d\") " pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.137912 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd9xn\" (UniqueName: \"kubernetes.io/projected/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-kube-api-access-bd9xn\") pod \"dnsmasq-dns-57d769cc4f-tf8jx\" (UID: \"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d\") " pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.311410 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.526425 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qcp96"] Dec 02 01:54:35 crc kubenswrapper[4884]: W1202 01:54:35.531976 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79c81e8b_e6ca_4f6b_b88c_9c9ebe5c4214.slice/crio-2a4632d7519c82ea0f2125d0c01ac4c6a8636068aa577e34023596a17506167b WatchSource:0}: Error finding container 2a4632d7519c82ea0f2125d0c01ac4c6a8636068aa577e34023596a17506167b: Status 404 returned error can't find the container with id 2a4632d7519c82ea0f2125d0c01ac4c6a8636068aa577e34023596a17506167b Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.610942 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tf8jx"] Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.823067 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.826922 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.829446 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.830847 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.831392 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.831411 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.831501 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-98v8z" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.832495 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.832844 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.835502 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.930732 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.930851 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.930897 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.930937 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-config-data\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.930960 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.930980 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.931010 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.931034 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm5x9\" (UniqueName: \"kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-kube-api-access-sm5x9\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.931058 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.931092 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f32d61be-992e-4a09-9663-33d3d59779e9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:35 crc kubenswrapper[4884]: I1202 01:54:35.931115 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f32d61be-992e-4a09-9663-33d3d59779e9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.032913 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.032962 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.033003 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-config-data\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.033024 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.033040 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.033062 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.033075 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm5x9\" (UniqueName: \"kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-kube-api-access-sm5x9\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.033099 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.033120 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f32d61be-992e-4a09-9663-33d3d59779e9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.033138 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f32d61be-992e-4a09-9663-33d3d59779e9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.033162 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.033508 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.033560 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.034065 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.034222 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.034423 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-config-data\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.035029 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.038262 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f32d61be-992e-4a09-9663-33d3d59779e9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.038428 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.038768 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.039046 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f32d61be-992e-4a09-9663-33d3d59779e9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.052706 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm5x9\" (UniqueName: \"kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-kube-api-access-sm5x9\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.056506 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.113929 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.115356 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.117682 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.117900 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.118436 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.118485 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.119047 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qhmqn" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.119302 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.120111 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.140057 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.163653 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.236506 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.236562 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kk59\" (UniqueName: \"kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-kube-api-access-7kk59\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.236613 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.236645 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.236674 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/27182112-178e-415c-9af9-f7fbf58d8120-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.236697 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.237383 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.237792 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.237820 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/27182112-178e-415c-9af9-f7fbf58d8120-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.237974 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.238285 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.339887 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.340170 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.340188 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/27182112-178e-415c-9af9-f7fbf58d8120-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.340211 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.340244 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.340266 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.340290 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kk59\" (UniqueName: \"kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-kube-api-access-7kk59\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.340307 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.340326 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.340348 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/27182112-178e-415c-9af9-f7fbf58d8120-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.340367 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.340826 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.341412 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.342614 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.342873 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.343008 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.343208 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.355981 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/27182112-178e-415c-9af9-f7fbf58d8120-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.356346 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.356476 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.356672 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/27182112-178e-415c-9af9-f7fbf58d8120-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.362987 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kk59\" (UniqueName: \"kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-kube-api-access-7kk59\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.366075 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.438984 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.506428 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qcp96" event={"ID":"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214","Type":"ContainerStarted","Data":"2a4632d7519c82ea0f2125d0c01ac4c6a8636068aa577e34023596a17506167b"} Dec 02 01:54:36 crc kubenswrapper[4884]: I1202 01:54:36.509453 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" event={"ID":"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d","Type":"ContainerStarted","Data":"9acc15b2d1dae6995338120a46845009c9a5df6327139342c5349e4a616bed31"} Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.554376 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.555655 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.557895 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.558241 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-4c5v6" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.558784 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.559047 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.565543 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.576850 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.662665 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/63a48d7f-b557-4f21-93d6-8a18e96eea16-config-data-generated\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.662714 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63a48d7f-b557-4f21-93d6-8a18e96eea16-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.662763 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/63a48d7f-b557-4f21-93d6-8a18e96eea16-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.662830 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/63a48d7f-b557-4f21-93d6-8a18e96eea16-kolla-config\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.662856 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/63a48d7f-b557-4f21-93d6-8a18e96eea16-config-data-default\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.662894 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b29bn\" (UniqueName: \"kubernetes.io/projected/63a48d7f-b557-4f21-93d6-8a18e96eea16-kube-api-access-b29bn\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.663031 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63a48d7f-b557-4f21-93d6-8a18e96eea16-operator-scripts\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.663080 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.764030 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63a48d7f-b557-4f21-93d6-8a18e96eea16-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.764085 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/63a48d7f-b557-4f21-93d6-8a18e96eea16-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.764140 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/63a48d7f-b557-4f21-93d6-8a18e96eea16-kolla-config\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.764174 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/63a48d7f-b557-4f21-93d6-8a18e96eea16-config-data-default\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.764652 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b29bn\" (UniqueName: \"kubernetes.io/projected/63a48d7f-b557-4f21-93d6-8a18e96eea16-kube-api-access-b29bn\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.764685 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63a48d7f-b557-4f21-93d6-8a18e96eea16-operator-scripts\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.764705 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.764802 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/63a48d7f-b557-4f21-93d6-8a18e96eea16-config-data-generated\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.765097 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/63a48d7f-b557-4f21-93d6-8a18e96eea16-config-data-generated\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.765183 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/63a48d7f-b557-4f21-93d6-8a18e96eea16-kolla-config\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.767806 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.767931 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/63a48d7f-b557-4f21-93d6-8a18e96eea16-config-data-default\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.768108 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63a48d7f-b557-4f21-93d6-8a18e96eea16-operator-scripts\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.785087 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/63a48d7f-b557-4f21-93d6-8a18e96eea16-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.785128 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63a48d7f-b557-4f21-93d6-8a18e96eea16-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.793997 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b29bn\" (UniqueName: \"kubernetes.io/projected/63a48d7f-b557-4f21-93d6-8a18e96eea16-kube-api-access-b29bn\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.795001 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-0\" (UID: \"63a48d7f-b557-4f21-93d6-8a18e96eea16\") " pod="openstack/openstack-galera-0" Dec 02 01:54:37 crc kubenswrapper[4884]: I1202 01:54:37.881247 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.194434 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.197437 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.205266 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.205484 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.205672 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-6psrg" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.206270 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.215669 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.288460 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.288515 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a8992d93-8785-4e52-af5b-73af8e0f56e5-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.288534 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a8992d93-8785-4e52-af5b-73af8e0f56e5-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.288553 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8992d93-8785-4e52-af5b-73af8e0f56e5-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.288576 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8992d93-8785-4e52-af5b-73af8e0f56e5-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.288590 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xn9qj\" (UniqueName: \"kubernetes.io/projected/a8992d93-8785-4e52-af5b-73af8e0f56e5-kube-api-access-xn9qj\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.288625 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a8992d93-8785-4e52-af5b-73af8e0f56e5-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.288655 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8992d93-8785-4e52-af5b-73af8e0f56e5-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.391574 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.391626 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a8992d93-8785-4e52-af5b-73af8e0f56e5-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.391658 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a8992d93-8785-4e52-af5b-73af8e0f56e5-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.391686 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8992d93-8785-4e52-af5b-73af8e0f56e5-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.391720 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8992d93-8785-4e52-af5b-73af8e0f56e5-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.391763 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xn9qj\" (UniqueName: \"kubernetes.io/projected/a8992d93-8785-4e52-af5b-73af8e0f56e5-kube-api-access-xn9qj\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.391809 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a8992d93-8785-4e52-af5b-73af8e0f56e5-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.391856 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8992d93-8785-4e52-af5b-73af8e0f56e5-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.392465 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a8992d93-8785-4e52-af5b-73af8e0f56e5-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.392734 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a8992d93-8785-4e52-af5b-73af8e0f56e5-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.392754 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a8992d93-8785-4e52-af5b-73af8e0f56e5-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.392954 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.393252 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a8992d93-8785-4e52-af5b-73af8e0f56e5-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.397174 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8992d93-8785-4e52-af5b-73af8e0f56e5-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.399719 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a8992d93-8785-4e52-af5b-73af8e0f56e5-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.423416 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xn9qj\" (UniqueName: \"kubernetes.io/projected/a8992d93-8785-4e52-af5b-73af8e0f56e5-kube-api-access-xn9qj\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.450122 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a8992d93-8785-4e52-af5b-73af8e0f56e5\") " pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.571598 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.572538 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.572672 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.578054 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.578152 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-4gvw6" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.578298 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.635383 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.695417 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dc21a47b-5514-4bd1-b164-05eb04fe2e31-kolla-config\") pod \"memcached-0\" (UID: \"dc21a47b-5514-4bd1-b164-05eb04fe2e31\") " pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.695526 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6d6f\" (UniqueName: \"kubernetes.io/projected/dc21a47b-5514-4bd1-b164-05eb04fe2e31-kube-api-access-h6d6f\") pod \"memcached-0\" (UID: \"dc21a47b-5514-4bd1-b164-05eb04fe2e31\") " pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.695735 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc21a47b-5514-4bd1-b164-05eb04fe2e31-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dc21a47b-5514-4bd1-b164-05eb04fe2e31\") " pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.696050 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc21a47b-5514-4bd1-b164-05eb04fe2e31-config-data\") pod \"memcached-0\" (UID: \"dc21a47b-5514-4bd1-b164-05eb04fe2e31\") " pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.696138 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc21a47b-5514-4bd1-b164-05eb04fe2e31-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dc21a47b-5514-4bd1-b164-05eb04fe2e31\") " pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.797968 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc21a47b-5514-4bd1-b164-05eb04fe2e31-config-data\") pod \"memcached-0\" (UID: \"dc21a47b-5514-4bd1-b164-05eb04fe2e31\") " pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.798026 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc21a47b-5514-4bd1-b164-05eb04fe2e31-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dc21a47b-5514-4bd1-b164-05eb04fe2e31\") " pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.798047 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dc21a47b-5514-4bd1-b164-05eb04fe2e31-kolla-config\") pod \"memcached-0\" (UID: \"dc21a47b-5514-4bd1-b164-05eb04fe2e31\") " pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.798082 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6d6f\" (UniqueName: \"kubernetes.io/projected/dc21a47b-5514-4bd1-b164-05eb04fe2e31-kube-api-access-h6d6f\") pod \"memcached-0\" (UID: \"dc21a47b-5514-4bd1-b164-05eb04fe2e31\") " pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.798114 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc21a47b-5514-4bd1-b164-05eb04fe2e31-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dc21a47b-5514-4bd1-b164-05eb04fe2e31\") " pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.798838 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc21a47b-5514-4bd1-b164-05eb04fe2e31-config-data\") pod \"memcached-0\" (UID: \"dc21a47b-5514-4bd1-b164-05eb04fe2e31\") " pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.798844 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dc21a47b-5514-4bd1-b164-05eb04fe2e31-kolla-config\") pod \"memcached-0\" (UID: \"dc21a47b-5514-4bd1-b164-05eb04fe2e31\") " pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.804551 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc21a47b-5514-4bd1-b164-05eb04fe2e31-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dc21a47b-5514-4bd1-b164-05eb04fe2e31\") " pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.809237 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc21a47b-5514-4bd1-b164-05eb04fe2e31-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dc21a47b-5514-4bd1-b164-05eb04fe2e31\") " pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.814466 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6d6f\" (UniqueName: \"kubernetes.io/projected/dc21a47b-5514-4bd1-b164-05eb04fe2e31-kube-api-access-h6d6f\") pod \"memcached-0\" (UID: \"dc21a47b-5514-4bd1-b164-05eb04fe2e31\") " pod="openstack/memcached-0" Dec 02 01:54:39 crc kubenswrapper[4884]: I1202 01:54:39.888127 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 02 01:54:41 crc kubenswrapper[4884]: I1202 01:54:41.590115 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 01:54:41 crc kubenswrapper[4884]: I1202 01:54:41.591809 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 01:54:41 crc kubenswrapper[4884]: I1202 01:54:41.593543 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-swj9z" Dec 02 01:54:41 crc kubenswrapper[4884]: I1202 01:54:41.606100 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 01:54:41 crc kubenswrapper[4884]: I1202 01:54:41.731532 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk2jc\" (UniqueName: \"kubernetes.io/projected/1583598a-ff88-4fea-bc25-5ff097443e7f-kube-api-access-mk2jc\") pod \"kube-state-metrics-0\" (UID: \"1583598a-ff88-4fea-bc25-5ff097443e7f\") " pod="openstack/kube-state-metrics-0" Dec 02 01:54:41 crc kubenswrapper[4884]: I1202 01:54:41.833514 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk2jc\" (UniqueName: \"kubernetes.io/projected/1583598a-ff88-4fea-bc25-5ff097443e7f-kube-api-access-mk2jc\") pod \"kube-state-metrics-0\" (UID: \"1583598a-ff88-4fea-bc25-5ff097443e7f\") " pod="openstack/kube-state-metrics-0" Dec 02 01:54:41 crc kubenswrapper[4884]: I1202 01:54:41.872815 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk2jc\" (UniqueName: \"kubernetes.io/projected/1583598a-ff88-4fea-bc25-5ff097443e7f-kube-api-access-mk2jc\") pod \"kube-state-metrics-0\" (UID: \"1583598a-ff88-4fea-bc25-5ff097443e7f\") " pod="openstack/kube-state-metrics-0" Dec 02 01:54:41 crc kubenswrapper[4884]: I1202 01:54:41.922326 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.856652 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.858473 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.862976 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.863060 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.863094 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.863164 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.863402 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-mcslp" Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.870957 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.871535 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.947784 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.947834 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.947855 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-config\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.947877 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.947907 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfstp\" (UniqueName: \"kubernetes.io/projected/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-kube-api-access-dfstp\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.947927 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.947968 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:42 crc kubenswrapper[4884]: I1202 01:54:42.948157 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.050919 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfstp\" (UniqueName: \"kubernetes.io/projected/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-kube-api-access-dfstp\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.051056 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.051103 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.051163 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.051251 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.051282 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.051307 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-config\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.051344 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.052144 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.056347 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.056956 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.057319 4884 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.057348 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1350ca4960854133c6cd5b98efbfa142de46bc322b82e157adbca982350a030b/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.057787 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.063968 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-config\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.077538 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfstp\" (UniqueName: \"kubernetes.io/projected/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-kube-api-access-dfstp\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.077589 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.096428 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") pod \"prometheus-metric-storage-0\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:43 crc kubenswrapper[4884]: I1202 01:54:43.187099 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.433959 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.436314 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.442481 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.447032 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-pddmq" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.447090 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.447199 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.447382 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.456096 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.607025 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/da645b5a-4282-4b4f-8eda-8713f990e51f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.607185 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/da645b5a-4282-4b4f-8eda-8713f990e51f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.607253 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/da645b5a-4282-4b4f-8eda-8713f990e51f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.607326 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.607396 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da645b5a-4282-4b4f-8eda-8713f990e51f-config\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.607517 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrjhw\" (UniqueName: \"kubernetes.io/projected/da645b5a-4282-4b4f-8eda-8713f990e51f-kube-api-access-vrjhw\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.607646 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da645b5a-4282-4b4f-8eda-8713f990e51f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.607812 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da645b5a-4282-4b4f-8eda-8713f990e51f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.708815 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da645b5a-4282-4b4f-8eda-8713f990e51f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.708874 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/da645b5a-4282-4b4f-8eda-8713f990e51f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.708923 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/da645b5a-4282-4b4f-8eda-8713f990e51f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.708948 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/da645b5a-4282-4b4f-8eda-8713f990e51f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.708980 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.709011 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da645b5a-4282-4b4f-8eda-8713f990e51f-config\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.709062 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrjhw\" (UniqueName: \"kubernetes.io/projected/da645b5a-4282-4b4f-8eda-8713f990e51f-kube-api-access-vrjhw\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.709107 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da645b5a-4282-4b4f-8eda-8713f990e51f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.710252 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.710572 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da645b5a-4282-4b4f-8eda-8713f990e51f-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.710658 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/da645b5a-4282-4b4f-8eda-8713f990e51f-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.711309 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da645b5a-4282-4b4f-8eda-8713f990e51f-config\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.715342 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/da645b5a-4282-4b4f-8eda-8713f990e51f-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.715407 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da645b5a-4282-4b4f-8eda-8713f990e51f-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.728403 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/da645b5a-4282-4b4f-8eda-8713f990e51f-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.728858 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrjhw\" (UniqueName: \"kubernetes.io/projected/da645b5a-4282-4b4f-8eda-8713f990e51f-kube-api-access-vrjhw\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.735545 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"da645b5a-4282-4b4f-8eda-8713f990e51f\") " pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:46 crc kubenswrapper[4884]: I1202 01:54:46.769435 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.173010 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-9wkcq"] Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.174229 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.176310 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.176618 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.183077 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-qfzp8" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.184121 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9wkcq"] Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.218689 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-bdv77"] Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.224897 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.236670 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-bdv77"] Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.319185 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nthfv\" (UniqueName: \"kubernetes.io/projected/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-kube-api-access-nthfv\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.319268 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-var-log-ovn\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.319319 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-ovn-controller-tls-certs\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.319370 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-scripts\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.320296 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-combined-ca-bundle\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.320398 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-var-run-ovn\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.320451 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-var-run\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.421917 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-var-log-ovn\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.421968 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-ovn-controller-tls-certs\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.422009 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-scripts\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.422044 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2536c28e-5917-4fc9-b37a-2f9212574dc5-etc-ovs\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.422069 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2536c28e-5917-4fc9-b37a-2f9212574dc5-scripts\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.422097 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-combined-ca-bundle\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.422124 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2536c28e-5917-4fc9-b37a-2f9212574dc5-var-lib\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.422149 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-var-run-ovn\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.422166 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2536c28e-5917-4fc9-b37a-2f9212574dc5-var-run\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.422188 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-var-run\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.422207 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2536c28e-5917-4fc9-b37a-2f9212574dc5-var-log\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.422233 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfh8v\" (UniqueName: \"kubernetes.io/projected/2536c28e-5917-4fc9-b37a-2f9212574dc5-kube-api-access-vfh8v\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.422255 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nthfv\" (UniqueName: \"kubernetes.io/projected/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-kube-api-access-nthfv\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.422865 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-var-log-ovn\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.425226 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-var-run-ovn\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.425385 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-var-run\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.428717 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-combined-ca-bundle\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.430003 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-scripts\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.431391 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-ovn-controller-tls-certs\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.437724 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nthfv\" (UniqueName: \"kubernetes.io/projected/cdabd8d6-7f1f-4629-b6cb-1be182bd9b66-kube-api-access-nthfv\") pod \"ovn-controller-9wkcq\" (UID: \"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66\") " pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.511500 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9wkcq" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.524354 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2536c28e-5917-4fc9-b37a-2f9212574dc5-var-run\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.524426 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2536c28e-5917-4fc9-b37a-2f9212574dc5-var-log\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.524459 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfh8v\" (UniqueName: \"kubernetes.io/projected/2536c28e-5917-4fc9-b37a-2f9212574dc5-kube-api-access-vfh8v\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.524556 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2536c28e-5917-4fc9-b37a-2f9212574dc5-etc-ovs\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.524586 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2536c28e-5917-4fc9-b37a-2f9212574dc5-scripts\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.524635 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2536c28e-5917-4fc9-b37a-2f9212574dc5-var-lib\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.524941 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2536c28e-5917-4fc9-b37a-2f9212574dc5-var-lib\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.525059 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2536c28e-5917-4fc9-b37a-2f9212574dc5-var-run\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.525121 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2536c28e-5917-4fc9-b37a-2f9212574dc5-etc-ovs\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.525232 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2536c28e-5917-4fc9-b37a-2f9212574dc5-var-log\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.527777 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2536c28e-5917-4fc9-b37a-2f9212574dc5-scripts\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.540345 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfh8v\" (UniqueName: \"kubernetes.io/projected/2536c28e-5917-4fc9-b37a-2f9212574dc5-kube-api-access-vfh8v\") pod \"ovn-controller-ovs-bdv77\" (UID: \"2536c28e-5917-4fc9-b37a-2f9212574dc5\") " pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:47 crc kubenswrapper[4884]: I1202 01:54:47.547077 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.167057 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.168496 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.170461 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.170767 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.171098 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.171263 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-8z2wj" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.178130 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.277539 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.277597 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.277619 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-config\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.277677 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9hj5\" (UniqueName: \"kubernetes.io/projected/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-kube-api-access-f9hj5\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.277819 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.277863 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.277884 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.278075 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.378784 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.378831 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.378858 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.378933 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.378987 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.379023 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.379175 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-config\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.379364 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.379429 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.380069 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-config\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.380265 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.380310 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9hj5\" (UniqueName: \"kubernetes.io/projected/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-kube-api-access-f9hj5\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.385256 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.386000 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.386526 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.403096 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.405337 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9hj5\" (UniqueName: \"kubernetes.io/projected/d16c1f3a-feb6-4809-b3d0-7306fb3d20ba-kube-api-access-f9hj5\") pod \"ovsdbserver-nb-0\" (UID: \"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba\") " pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:49 crc kubenswrapper[4884]: I1202 01:54:49.545173 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 02 01:54:50 crc kubenswrapper[4884]: E1202 01:54:50.306302 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 02 01:54:50 crc kubenswrapper[4884]: E1202 01:54:50.306529 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cj9fh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-wf8w2_openstack(ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 01:54:50 crc kubenswrapper[4884]: E1202 01:54:50.307952 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" podUID="ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968" Dec 02 01:54:50 crc kubenswrapper[4884]: E1202 01:54:50.321457 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 02 01:54:50 crc kubenswrapper[4884]: E1202 01:54:50.321615 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tk7kx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-6zx6q_openstack(4c1ac317-3e2a-435c-bd07-933cb4e4fcc3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 01:54:50 crc kubenswrapper[4884]: E1202 01:54:50.322682 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-6zx6q" podUID="4c1ac317-3e2a-435c-bd07-933cb4e4fcc3" Dec 02 01:54:50 crc kubenswrapper[4884]: I1202 01:54:50.850851 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.129069 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.152022 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.270266 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-6zx6q" Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.311527 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c1ac317-3e2a-435c-bd07-933cb4e4fcc3-config\") pod \"4c1ac317-3e2a-435c-bd07-933cb4e4fcc3\" (UID: \"4c1ac317-3e2a-435c-bd07-933cb4e4fcc3\") " Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.311592 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk7kx\" (UniqueName: \"kubernetes.io/projected/4c1ac317-3e2a-435c-bd07-933cb4e4fcc3-kube-api-access-tk7kx\") pod \"4c1ac317-3e2a-435c-bd07-933cb4e4fcc3\" (UID: \"4c1ac317-3e2a-435c-bd07-933cb4e4fcc3\") " Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.313029 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c1ac317-3e2a-435c-bd07-933cb4e4fcc3-config" (OuterVolumeSpecName: "config") pod "4c1ac317-3e2a-435c-bd07-933cb4e4fcc3" (UID: "4c1ac317-3e2a-435c-bd07-933cb4e4fcc3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.316946 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c1ac317-3e2a-435c-bd07-933cb4e4fcc3-kube-api-access-tk7kx" (OuterVolumeSpecName: "kube-api-access-tk7kx") pod "4c1ac317-3e2a-435c-bd07-933cb4e4fcc3" (UID: "4c1ac317-3e2a-435c-bd07-933cb4e4fcc3"). InnerVolumeSpecName "kube-api-access-tk7kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.426085 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk7kx\" (UniqueName: \"kubernetes.io/projected/4c1ac317-3e2a-435c-bd07-933cb4e4fcc3-kube-api-access-tk7kx\") on node \"crc\" DevicePath \"\"" Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.426162 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c1ac317-3e2a-435c-bd07-933cb4e4fcc3-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.478051 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 01:54:51 crc kubenswrapper[4884]: W1202 01:54:51.478737 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1583598a_ff88_4fea_bc25_5ff097443e7f.slice/crio-7259233685c2d6fd6d95e04abfe630073fe780410305d7d91fcc4bd751a91e67 WatchSource:0}: Error finding container 7259233685c2d6fd6d95e04abfe630073fe780410305d7d91fcc4bd751a91e67: Status 404 returned error can't find the container with id 7259233685c2d6fd6d95e04abfe630073fe780410305d7d91fcc4bd751a91e67 Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.489808 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 02 01:54:51 crc kubenswrapper[4884]: W1202 01:54:51.490778 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdabd8d6_7f1f_4629_b6cb_1be182bd9b66.slice/crio-de68315f907903f0416d493ba9a5eb3851fe060d6f01b68830c1c1ed25eed89a WatchSource:0}: Error finding container de68315f907903f0416d493ba9a5eb3851fe060d6f01b68830c1c1ed25eed89a: Status 404 returned error can't find the container with id de68315f907903f0416d493ba9a5eb3851fe060d6f01b68830c1c1ed25eed89a Dec 02 01:54:51 crc kubenswrapper[4884]: W1202 01:54:51.492639 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc21a47b_5514_4bd1_b164_05eb04fe2e31.slice/crio-932e229849d4822001d023c75dc1dcbefe3adde21b00e1cb0a972e2145940f10 WatchSource:0}: Error finding container 932e229849d4822001d023c75dc1dcbefe3adde21b00e1cb0a972e2145940f10: Status 404 returned error can't find the container with id 932e229849d4822001d023c75dc1dcbefe3adde21b00e1cb0a972e2145940f10 Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.494704 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9wkcq"] Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.661061 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.702302 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.717186 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 02 01:54:51 crc kubenswrapper[4884]: W1202 01:54:51.795103 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd16c1f3a_feb6_4809_b3d0_7306fb3d20ba.slice/crio-4c424730b79244c4129b8795ca45b39dcd4ff82e313472e02c7810315ddac7fb WatchSource:0}: Error finding container 4c424730b79244c4129b8795ca45b39dcd4ff82e313472e02c7810315ddac7fb: Status 404 returned error can't find the container with id 4c424730b79244c4129b8795ca45b39dcd4ff82e313472e02c7810315ddac7fb Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.795128 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.842047 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1583598a-ff88-4fea-bc25-5ff097443e7f","Type":"ContainerStarted","Data":"7259233685c2d6fd6d95e04abfe630073fe780410305d7d91fcc4bd751a91e67"} Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.843205 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-config\") pod \"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968\" (UID: \"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968\") " Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.843325 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj9fh\" (UniqueName: \"kubernetes.io/projected/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-kube-api-access-cj9fh\") pod \"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968\" (UID: \"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968\") " Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.843415 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-dns-svc\") pod \"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968\" (UID: \"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968\") " Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.843631 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9wkcq" event={"ID":"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66","Type":"ContainerStarted","Data":"de68315f907903f0416d493ba9a5eb3851fe060d6f01b68830c1c1ed25eed89a"} Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.843779 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-config" (OuterVolumeSpecName: "config") pod "ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968" (UID: "ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.843881 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.844377 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968" (UID: "ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.845784 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a8992d93-8785-4e52-af5b-73af8e0f56e5","Type":"ContainerStarted","Data":"d5158054637db6ff8c35ef3285811ed22f1c64251d61233e502233478473bfec"} Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.847798 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-kube-api-access-cj9fh" (OuterVolumeSpecName: "kube-api-access-cj9fh") pod "ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968" (UID: "ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968"). InnerVolumeSpecName "kube-api-access-cj9fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.848081 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.848113 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-wf8w2" event={"ID":"ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968","Type":"ContainerDied","Data":"eaeb9f8e69c5a405a8e5e7b7f1f54d47684849a126e6158b9d0aa425580d866c"} Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.849486 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-6zx6q" event={"ID":"4c1ac317-3e2a-435c-bd07-933cb4e4fcc3","Type":"ContainerDied","Data":"826112fbf353ff452a4e445197d3ee5a6222fdfd6aa4e4d10827f7f0e40761da"} Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.849532 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-6zx6q" Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.851487 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"dc21a47b-5514-4bd1-b164-05eb04fe2e31","Type":"ContainerStarted","Data":"932e229849d4822001d023c75dc1dcbefe3adde21b00e1cb0a972e2145940f10"} Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.853820 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f32d61be-992e-4a09-9663-33d3d59779e9","Type":"ContainerStarted","Data":"cd0bcd038f0ce6837b08f6f928b5bd49d8ea3667fb693985e21a97472f5f63b2"} Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.855668 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"27182112-178e-415c-9af9-f7fbf58d8120","Type":"ContainerStarted","Data":"448febc64be0ef9b1b26959f87b8bfd21fe6ef4b9a57b1cc6c5804bb801e22b2"} Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.857209 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c8bb2562-5d06-44d1-bbd0-8848d92f18d2","Type":"ContainerStarted","Data":"812130cbe2c9e5d1331a385a71c447a151390e2b49670622b49eeabdaabdc691"} Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.861368 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba","Type":"ContainerStarted","Data":"4c424730b79244c4129b8795ca45b39dcd4ff82e313472e02c7810315ddac7fb"} Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.866885 4884 generic.go:334] "Generic (PLEG): container finished" podID="01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d" containerID="27ba7839a91e5ed64eb8095c1c7248d8a230bc2665f1b4ffeab8f98debcd9c8a" exitCode=0 Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.866948 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" event={"ID":"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d","Type":"ContainerDied","Data":"27ba7839a91e5ed64eb8095c1c7248d8a230bc2665f1b4ffeab8f98debcd9c8a"} Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.890717 4884 generic.go:334] "Generic (PLEG): container finished" podID="79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214" containerID="be3d4b78b1a32165cff51230d3c8c7481dc21e37349c3f42780adcc101ef518b" exitCode=0 Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.890851 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qcp96" event={"ID":"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214","Type":"ContainerDied","Data":"be3d4b78b1a32165cff51230d3c8c7481dc21e37349c3f42780adcc101ef518b"} Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.895956 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6zx6q"] Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.897834 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"63a48d7f-b557-4f21-93d6-8a18e96eea16","Type":"ContainerStarted","Data":"049fc677c406fe843fd2d7acb8352705d9c23d4c818a4ac9b62d653f5a6168c9"} Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.907004 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-6zx6q"] Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.921335 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-bdv77"] Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.941871 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wf8w2"] Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.944863 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj9fh\" (UniqueName: \"kubernetes.io/projected/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-kube-api-access-cj9fh\") on node \"crc\" DevicePath \"\"" Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.944889 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 01:54:51 crc kubenswrapper[4884]: I1202 01:54:51.951970 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wf8w2"] Dec 02 01:54:52 crc kubenswrapper[4884]: E1202 01:54:52.219780 4884 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 02 01:54:52 crc kubenswrapper[4884]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 02 01:54:52 crc kubenswrapper[4884]: > podSandboxID="2a4632d7519c82ea0f2125d0c01ac4c6a8636068aa577e34023596a17506167b" Dec 02 01:54:52 crc kubenswrapper[4884]: E1202 01:54:52.220172 4884 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 02 01:54:52 crc kubenswrapper[4884]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-df8fd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-qcp96_openstack(79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 02 01:54:52 crc kubenswrapper[4884]: > logger="UnhandledError" Dec 02 01:54:52 crc kubenswrapper[4884]: E1202 01:54:52.221326 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-666b6646f7-qcp96" podUID="79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214" Dec 02 01:54:52 crc kubenswrapper[4884]: I1202 01:54:52.730359 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 02 01:54:52 crc kubenswrapper[4884]: I1202 01:54:52.906762 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" event={"ID":"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d","Type":"ContainerStarted","Data":"256f1c2d98576266fed9b5f4578d9424e93076aeaf6f42686017f7323a6dfa1f"} Dec 02 01:54:52 crc kubenswrapper[4884]: I1202 01:54:52.907674 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" Dec 02 01:54:52 crc kubenswrapper[4884]: I1202 01:54:52.909290 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bdv77" event={"ID":"2536c28e-5917-4fc9-b37a-2f9212574dc5","Type":"ContainerStarted","Data":"cbf4b126b7994bb4a364c971167886f913863a18e7a13d74076c2ecc25bd0c13"} Dec 02 01:54:52 crc kubenswrapper[4884]: I1202 01:54:52.928031 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" podStartSLOduration=3.952009344 podStartE2EDuration="18.928015586s" podCreationTimestamp="2025-12-02 01:54:34 +0000 UTC" firstStartedPulling="2025-12-02 01:54:35.629613357 +0000 UTC m=+1092.305450241" lastFinishedPulling="2025-12-02 01:54:50.605619599 +0000 UTC m=+1107.281456483" observedRunningTime="2025-12-02 01:54:52.927374639 +0000 UTC m=+1109.603211533" watchObservedRunningTime="2025-12-02 01:54:52.928015586 +0000 UTC m=+1109.603852470" Dec 02 01:54:53 crc kubenswrapper[4884]: W1202 01:54:53.541835 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda645b5a_4282_4b4f_8eda_8713f990e51f.slice/crio-f4243310a2c5f75bf49864720f102ffcf95714f526fe9cb7d6c383e24a823b3d WatchSource:0}: Error finding container f4243310a2c5f75bf49864720f102ffcf95714f526fe9cb7d6c383e24a823b3d: Status 404 returned error can't find the container with id f4243310a2c5f75bf49864720f102ffcf95714f526fe9cb7d6c383e24a823b3d Dec 02 01:54:53 crc kubenswrapper[4884]: I1202 01:54:53.628282 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c1ac317-3e2a-435c-bd07-933cb4e4fcc3" path="/var/lib/kubelet/pods/4c1ac317-3e2a-435c-bd07-933cb4e4fcc3/volumes" Dec 02 01:54:53 crc kubenswrapper[4884]: I1202 01:54:53.628654 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968" path="/var/lib/kubelet/pods/ac3a8976-c9ea-4ec8-beb7-b1bfa90a8968/volumes" Dec 02 01:54:53 crc kubenswrapper[4884]: I1202 01:54:53.916106 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"da645b5a-4282-4b4f-8eda-8713f990e51f","Type":"ContainerStarted","Data":"f4243310a2c5f75bf49864720f102ffcf95714f526fe9cb7d6c383e24a823b3d"} Dec 02 01:55:00 crc kubenswrapper[4884]: I1202 01:55:00.312925 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" Dec 02 01:55:00 crc kubenswrapper[4884]: I1202 01:55:00.381839 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qcp96"] Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.015363 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-9wkcq" Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.018123 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9wkcq" event={"ID":"cdabd8d6-7f1f-4629-b6cb-1be182bd9b66","Type":"ContainerStarted","Data":"0883872a975a06048a7b7ebe0781bc3bc3eb95c518b446bf2c5e26ca33eb9b3c"} Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.022236 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"da645b5a-4282-4b4f-8eda-8713f990e51f","Type":"ContainerStarted","Data":"13b7c5936ebc8d57585e8d06c0d989cbcd81b08278ab5ee4b14c64134ff686d7"} Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.024860 4884 generic.go:334] "Generic (PLEG): container finished" podID="2536c28e-5917-4fc9-b37a-2f9212574dc5" containerID="72b4e708a96719dac569a8d844d8758db0f9c0e0b0a97178b2ef17e27f3a6f16" exitCode=0 Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.024921 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bdv77" event={"ID":"2536c28e-5917-4fc9-b37a-2f9212574dc5","Type":"ContainerDied","Data":"72b4e708a96719dac569a8d844d8758db0f9c0e0b0a97178b2ef17e27f3a6f16"} Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.031298 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qcp96" event={"ID":"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214","Type":"ContainerStarted","Data":"d18e01fde1aada1c86b364b05b48766ffafaf7f99627901f1ead0ca289e69c36"} Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.031475 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-qcp96" podUID="79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214" containerName="dnsmasq-dns" containerID="cri-o://d18e01fde1aada1c86b364b05b48766ffafaf7f99627901f1ead0ca289e69c36" gracePeriod=10 Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.031567 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-qcp96" Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.036056 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-9wkcq" podStartSLOduration=6.220191726 podStartE2EDuration="14.036034206s" podCreationTimestamp="2025-12-02 01:54:47 +0000 UTC" firstStartedPulling="2025-12-02 01:54:51.492212648 +0000 UTC m=+1108.168049532" lastFinishedPulling="2025-12-02 01:54:59.308055128 +0000 UTC m=+1115.983892012" observedRunningTime="2025-12-02 01:55:01.035577165 +0000 UTC m=+1117.711414059" watchObservedRunningTime="2025-12-02 01:55:01.036034206 +0000 UTC m=+1117.711871100" Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.040715 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba","Type":"ContainerStarted","Data":"6b23bd6cd1388429da7b966c6656f664bc446d496bfcb5dfbdd0bad7b4f8e8f2"} Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.044436 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"dc21a47b-5514-4bd1-b164-05eb04fe2e31","Type":"ContainerStarted","Data":"dfdefc0f2417279971e0411201d37a074fe1c9d2315e2595340f969e69c2a447"} Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.044610 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.052067 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1583598a-ff88-4fea-bc25-5ff097443e7f","Type":"ContainerStarted","Data":"f0935a91358e557e628541d3800d9ca9311441f0dd6c3fad9a949bfe6b0d8921"} Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.053079 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.057659 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a8992d93-8785-4e52-af5b-73af8e0f56e5","Type":"ContainerStarted","Data":"5611615a339c3ce4b0539c33e2314c2785ef8bfc1aae2bf3b56a2fd5930a7430"} Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.060158 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"63a48d7f-b557-4f21-93d6-8a18e96eea16","Type":"ContainerStarted","Data":"99818784aacc434bd5f2a051332e7dfe1c55ca39ac59f23a445d0b0462337b3b"} Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.065521 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-qcp96" podStartSLOduration=12.003073608 podStartE2EDuration="27.065506538s" podCreationTimestamp="2025-12-02 01:54:34 +0000 UTC" firstStartedPulling="2025-12-02 01:54:35.534428099 +0000 UTC m=+1092.210264983" lastFinishedPulling="2025-12-02 01:54:50.596861039 +0000 UTC m=+1107.272697913" observedRunningTime="2025-12-02 01:55:01.052229234 +0000 UTC m=+1117.728066138" watchObservedRunningTime="2025-12-02 01:55:01.065506538 +0000 UTC m=+1117.741343412" Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.090635 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=11.562604691 podStartE2EDuration="20.09061979s" podCreationTimestamp="2025-12-02 01:54:41 +0000 UTC" firstStartedPulling="2025-12-02 01:54:51.48115315 +0000 UTC m=+1108.156990034" lastFinishedPulling="2025-12-02 01:55:00.009168219 +0000 UTC m=+1116.685005133" observedRunningTime="2025-12-02 01:55:01.084550647 +0000 UTC m=+1117.760387531" watchObservedRunningTime="2025-12-02 01:55:01.09061979 +0000 UTC m=+1117.766456674" Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.127367 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.937515965 podStartE2EDuration="22.127347274s" podCreationTimestamp="2025-12-02 01:54:39 +0000 UTC" firstStartedPulling="2025-12-02 01:54:51.494608508 +0000 UTC m=+1108.170445402" lastFinishedPulling="2025-12-02 01:54:58.684439827 +0000 UTC m=+1115.360276711" observedRunningTime="2025-12-02 01:55:01.120694356 +0000 UTC m=+1117.796531240" watchObservedRunningTime="2025-12-02 01:55:01.127347274 +0000 UTC m=+1117.803184158" Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.542947 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qcp96" Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.701145 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-dns-svc\") pod \"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214\" (UID: \"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214\") " Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.702731 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-config\") pod \"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214\" (UID: \"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214\") " Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.702809 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df8fd\" (UniqueName: \"kubernetes.io/projected/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-kube-api-access-df8fd\") pod \"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214\" (UID: \"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214\") " Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.750191 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-kube-api-access-df8fd" (OuterVolumeSpecName: "kube-api-access-df8fd") pod "79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214" (UID: "79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214"). InnerVolumeSpecName "kube-api-access-df8fd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.811507 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df8fd\" (UniqueName: \"kubernetes.io/projected/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-kube-api-access-df8fd\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.869273 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-config" (OuterVolumeSpecName: "config") pod "79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214" (UID: "79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.881789 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214" (UID: "79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.913818 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:01 crc kubenswrapper[4884]: I1202 01:55:01.913856 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:02 crc kubenswrapper[4884]: I1202 01:55:02.068721 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f32d61be-992e-4a09-9663-33d3d59779e9","Type":"ContainerStarted","Data":"39010f45dce6a3ffe62ec61181fd5ae24a376cc18ac028cb8715cf1211681fc2"} Dec 02 01:55:02 crc kubenswrapper[4884]: I1202 01:55:02.070463 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"27182112-178e-415c-9af9-f7fbf58d8120","Type":"ContainerStarted","Data":"b9dfb8cc8de08b7d246a2396788751343d208c51ed83bbee65b5010ac935f251"} Dec 02 01:55:02 crc kubenswrapper[4884]: I1202 01:55:02.072774 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bdv77" event={"ID":"2536c28e-5917-4fc9-b37a-2f9212574dc5","Type":"ContainerStarted","Data":"88f754b5be7b22d141e03042334805e94ded548b62422967cd2b96022805241d"} Dec 02 01:55:02 crc kubenswrapper[4884]: I1202 01:55:02.074464 4884 generic.go:334] "Generic (PLEG): container finished" podID="79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214" containerID="d18e01fde1aada1c86b364b05b48766ffafaf7f99627901f1ead0ca289e69c36" exitCode=0 Dec 02 01:55:02 crc kubenswrapper[4884]: I1202 01:55:02.074563 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qcp96" event={"ID":"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214","Type":"ContainerDied","Data":"d18e01fde1aada1c86b364b05b48766ffafaf7f99627901f1ead0ca289e69c36"} Dec 02 01:55:02 crc kubenswrapper[4884]: I1202 01:55:02.074626 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-qcp96" event={"ID":"79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214","Type":"ContainerDied","Data":"2a4632d7519c82ea0f2125d0c01ac4c6a8636068aa577e34023596a17506167b"} Dec 02 01:55:02 crc kubenswrapper[4884]: I1202 01:55:02.074660 4884 scope.go:117] "RemoveContainer" containerID="d18e01fde1aada1c86b364b05b48766ffafaf7f99627901f1ead0ca289e69c36" Dec 02 01:55:02 crc kubenswrapper[4884]: I1202 01:55:02.074488 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-qcp96" Dec 02 01:55:02 crc kubenswrapper[4884]: I1202 01:55:02.302988 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qcp96"] Dec 02 01:55:02 crc kubenswrapper[4884]: I1202 01:55:02.312415 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-qcp96"] Dec 02 01:55:03 crc kubenswrapper[4884]: I1202 01:55:03.088571 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bdv77" event={"ID":"2536c28e-5917-4fc9-b37a-2f9212574dc5","Type":"ContainerStarted","Data":"cb6c37c18ccda60590a5d9f9a6021b88b1a42f186ad5f0e44d137035ccb4848c"} Dec 02 01:55:03 crc kubenswrapper[4884]: I1202 01:55:03.088809 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:55:03 crc kubenswrapper[4884]: I1202 01:55:03.091618 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c8bb2562-5d06-44d1-bbd0-8848d92f18d2","Type":"ContainerStarted","Data":"f632bbc70feae9bcab6e61183a7487126f6504963a2bb6247f36891b3ef48634"} Dec 02 01:55:03 crc kubenswrapper[4884]: I1202 01:55:03.108394 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-bdv77" podStartSLOduration=8.724045847 podStartE2EDuration="16.10836902s" podCreationTimestamp="2025-12-02 01:54:47 +0000 UTC" firstStartedPulling="2025-12-02 01:54:51.923677824 +0000 UTC m=+1108.599514708" lastFinishedPulling="2025-12-02 01:54:59.308000967 +0000 UTC m=+1115.983837881" observedRunningTime="2025-12-02 01:55:03.10755842 +0000 UTC m=+1119.783395314" watchObservedRunningTime="2025-12-02 01:55:03.10836902 +0000 UTC m=+1119.784205934" Dec 02 01:55:03 crc kubenswrapper[4884]: I1202 01:55:03.631818 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214" path="/var/lib/kubelet/pods/79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214/volumes" Dec 02 01:55:04 crc kubenswrapper[4884]: I1202 01:55:04.099901 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:55:04 crc kubenswrapper[4884]: I1202 01:55:04.451079 4884 scope.go:117] "RemoveContainer" containerID="be3d4b78b1a32165cff51230d3c8c7481dc21e37349c3f42780adcc101ef518b" Dec 02 01:55:04 crc kubenswrapper[4884]: I1202 01:55:04.984454 4884 scope.go:117] "RemoveContainer" containerID="d18e01fde1aada1c86b364b05b48766ffafaf7f99627901f1ead0ca289e69c36" Dec 02 01:55:04 crc kubenswrapper[4884]: E1202 01:55:04.986214 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d18e01fde1aada1c86b364b05b48766ffafaf7f99627901f1ead0ca289e69c36\": container with ID starting with d18e01fde1aada1c86b364b05b48766ffafaf7f99627901f1ead0ca289e69c36 not found: ID does not exist" containerID="d18e01fde1aada1c86b364b05b48766ffafaf7f99627901f1ead0ca289e69c36" Dec 02 01:55:04 crc kubenswrapper[4884]: I1202 01:55:04.986250 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d18e01fde1aada1c86b364b05b48766ffafaf7f99627901f1ead0ca289e69c36"} err="failed to get container status \"d18e01fde1aada1c86b364b05b48766ffafaf7f99627901f1ead0ca289e69c36\": rpc error: code = NotFound desc = could not find container \"d18e01fde1aada1c86b364b05b48766ffafaf7f99627901f1ead0ca289e69c36\": container with ID starting with d18e01fde1aada1c86b364b05b48766ffafaf7f99627901f1ead0ca289e69c36 not found: ID does not exist" Dec 02 01:55:04 crc kubenswrapper[4884]: I1202 01:55:04.986269 4884 scope.go:117] "RemoveContainer" containerID="be3d4b78b1a32165cff51230d3c8c7481dc21e37349c3f42780adcc101ef518b" Dec 02 01:55:04 crc kubenswrapper[4884]: E1202 01:55:04.986582 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be3d4b78b1a32165cff51230d3c8c7481dc21e37349c3f42780adcc101ef518b\": container with ID starting with be3d4b78b1a32165cff51230d3c8c7481dc21e37349c3f42780adcc101ef518b not found: ID does not exist" containerID="be3d4b78b1a32165cff51230d3c8c7481dc21e37349c3f42780adcc101ef518b" Dec 02 01:55:04 crc kubenswrapper[4884]: I1202 01:55:04.986625 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be3d4b78b1a32165cff51230d3c8c7481dc21e37349c3f42780adcc101ef518b"} err="failed to get container status \"be3d4b78b1a32165cff51230d3c8c7481dc21e37349c3f42780adcc101ef518b\": rpc error: code = NotFound desc = could not find container \"be3d4b78b1a32165cff51230d3c8c7481dc21e37349c3f42780adcc101ef518b\": container with ID starting with be3d4b78b1a32165cff51230d3c8c7481dc21e37349c3f42780adcc101ef518b not found: ID does not exist" Dec 02 01:55:05 crc kubenswrapper[4884]: I1202 01:55:05.107543 4884 generic.go:334] "Generic (PLEG): container finished" podID="a8992d93-8785-4e52-af5b-73af8e0f56e5" containerID="5611615a339c3ce4b0539c33e2314c2785ef8bfc1aae2bf3b56a2fd5930a7430" exitCode=0 Dec 02 01:55:05 crc kubenswrapper[4884]: I1202 01:55:05.107613 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a8992d93-8785-4e52-af5b-73af8e0f56e5","Type":"ContainerDied","Data":"5611615a339c3ce4b0539c33e2314c2785ef8bfc1aae2bf3b56a2fd5930a7430"} Dec 02 01:55:05 crc kubenswrapper[4884]: I1202 01:55:05.112922 4884 generic.go:334] "Generic (PLEG): container finished" podID="63a48d7f-b557-4f21-93d6-8a18e96eea16" containerID="99818784aacc434bd5f2a051332e7dfe1c55ca39ac59f23a445d0b0462337b3b" exitCode=0 Dec 02 01:55:05 crc kubenswrapper[4884]: I1202 01:55:05.113085 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"63a48d7f-b557-4f21-93d6-8a18e96eea16","Type":"ContainerDied","Data":"99818784aacc434bd5f2a051332e7dfe1c55ca39ac59f23a445d0b0462337b3b"} Dec 02 01:55:06 crc kubenswrapper[4884]: I1202 01:55:06.123176 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"63a48d7f-b557-4f21-93d6-8a18e96eea16","Type":"ContainerStarted","Data":"99787fef5141ba8a27fb09475934d98aee5ab43933af3682e7e7401a79792ca9"} Dec 02 01:55:06 crc kubenswrapper[4884]: I1202 01:55:06.125029 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"d16c1f3a-feb6-4809-b3d0-7306fb3d20ba","Type":"ContainerStarted","Data":"19395a4a029473a20933dd15db4a4c6f9a93352cc81dcae1d632f985d70cbcd9"} Dec 02 01:55:06 crc kubenswrapper[4884]: I1202 01:55:06.126826 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a8992d93-8785-4e52-af5b-73af8e0f56e5","Type":"ContainerStarted","Data":"18713163a3c09bb7ea0b2f914145ed0adda9000b6e79976b1543e96939eb5f58"} Dec 02 01:55:06 crc kubenswrapper[4884]: I1202 01:55:06.128698 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"da645b5a-4282-4b4f-8eda-8713f990e51f","Type":"ContainerStarted","Data":"defac0df9c3fe6ff5569b0d62b25cb266b2206b20978de01bd9a56b92f13dff5"} Dec 02 01:55:06 crc kubenswrapper[4884]: I1202 01:55:06.161144 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=22.572499471 podStartE2EDuration="30.161128143s" podCreationTimestamp="2025-12-02 01:54:36 +0000 UTC" firstStartedPulling="2025-12-02 01:54:51.719404685 +0000 UTC m=+1108.395241569" lastFinishedPulling="2025-12-02 01:54:59.308033327 +0000 UTC m=+1115.983870241" observedRunningTime="2025-12-02 01:55:06.156333722 +0000 UTC m=+1122.832170626" watchObservedRunningTime="2025-12-02 01:55:06.161128143 +0000 UTC m=+1122.836965027" Dec 02 01:55:06 crc kubenswrapper[4884]: I1202 01:55:06.191305 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=4.953315343 podStartE2EDuration="18.191287142s" podCreationTimestamp="2025-12-02 01:54:48 +0000 UTC" firstStartedPulling="2025-12-02 01:54:51.797303765 +0000 UTC m=+1108.473140669" lastFinishedPulling="2025-12-02 01:55:05.035275584 +0000 UTC m=+1121.711112468" observedRunningTime="2025-12-02 01:55:06.190092241 +0000 UTC m=+1122.865929125" watchObservedRunningTime="2025-12-02 01:55:06.191287142 +0000 UTC m=+1122.867124026" Dec 02 01:55:06 crc kubenswrapper[4884]: I1202 01:55:06.222309 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=9.73844569 podStartE2EDuration="21.222289262s" podCreationTimestamp="2025-12-02 01:54:45 +0000 UTC" firstStartedPulling="2025-12-02 01:54:53.544152179 +0000 UTC m=+1110.219989063" lastFinishedPulling="2025-12-02 01:55:05.027995741 +0000 UTC m=+1121.703832635" observedRunningTime="2025-12-02 01:55:06.214822764 +0000 UTC m=+1122.890659648" watchObservedRunningTime="2025-12-02 01:55:06.222289262 +0000 UTC m=+1122.898126136" Dec 02 01:55:06 crc kubenswrapper[4884]: I1202 01:55:06.242264 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=19.966050871 podStartE2EDuration="28.242247334s" podCreationTimestamp="2025-12-02 01:54:38 +0000 UTC" firstStartedPulling="2025-12-02 01:54:51.166702088 +0000 UTC m=+1107.842538972" lastFinishedPulling="2025-12-02 01:54:59.442898551 +0000 UTC m=+1116.118735435" observedRunningTime="2025-12-02 01:55:06.234246033 +0000 UTC m=+1122.910082927" watchObservedRunningTime="2025-12-02 01:55:06.242247334 +0000 UTC m=+1122.918084218" Dec 02 01:55:06 crc kubenswrapper[4884]: I1202 01:55:06.770246 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 02 01:55:07 crc kubenswrapper[4884]: I1202 01:55:07.545922 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 02 01:55:07 crc kubenswrapper[4884]: I1202 01:55:07.587250 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 02 01:55:07 crc kubenswrapper[4884]: I1202 01:55:07.770118 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 02 01:55:07 crc kubenswrapper[4884]: I1202 01:55:07.809051 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 02 01:55:07 crc kubenswrapper[4884]: I1202 01:55:07.882399 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 02 01:55:07 crc kubenswrapper[4884]: I1202 01:55:07.882480 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.149706 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.186936 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.231130 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.411495 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kvm9n"] Dec 02 01:55:08 crc kubenswrapper[4884]: E1202 01:55:08.411831 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214" containerName="init" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.411842 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214" containerName="init" Dec 02 01:55:08 crc kubenswrapper[4884]: E1202 01:55:08.411853 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214" containerName="dnsmasq-dns" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.411859 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214" containerName="dnsmasq-dns" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.412004 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="79c81e8b-e6ca-4f6b-b88c-9c9ebe5c4214" containerName="dnsmasq-dns" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.412790 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.418060 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.426085 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kvm9n"] Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.438688 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-kvm9n\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.438775 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dtx5\" (UniqueName: \"kubernetes.io/projected/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-kube-api-access-8dtx5\") pod \"dnsmasq-dns-7fd796d7df-kvm9n\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.438839 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-kvm9n\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.438857 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-config\") pod \"dnsmasq-dns-7fd796d7df-kvm9n\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.459334 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-hplrx"] Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.460685 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.464458 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.477303 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-hplrx"] Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.540692 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dtx5\" (UniqueName: \"kubernetes.io/projected/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-kube-api-access-8dtx5\") pod \"dnsmasq-dns-7fd796d7df-kvm9n\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.540770 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc44c0f6-5201-4b7a-85e9-db2a1766c237-config\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.540801 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc44c0f6-5201-4b7a-85e9-db2a1766c237-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.540845 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc44c0f6-5201-4b7a-85e9-db2a1766c237-combined-ca-bundle\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.540904 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-kvm9n\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.540922 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-config\") pod \"dnsmasq-dns-7fd796d7df-kvm9n\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.540980 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t2k5\" (UniqueName: \"kubernetes.io/projected/dc44c0f6-5201-4b7a-85e9-db2a1766c237-kube-api-access-2t2k5\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.540997 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-kvm9n\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.541013 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/dc44c0f6-5201-4b7a-85e9-db2a1766c237-ovs-rundir\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.541030 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/dc44c0f6-5201-4b7a-85e9-db2a1766c237-ovn-rundir\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.542160 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-kvm9n\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.542314 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-config\") pod \"dnsmasq-dns-7fd796d7df-kvm9n\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.542702 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-kvm9n\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.560047 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kvm9n"] Dec 02 01:55:08 crc kubenswrapper[4884]: E1202 01:55:08.560671 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-8dtx5], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" podUID="bd50d32a-dcda-4403-aff4-0eabdbcaa6d8" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.578095 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dtx5\" (UniqueName: \"kubernetes.io/projected/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-kube-api-access-8dtx5\") pod \"dnsmasq-dns-7fd796d7df-kvm9n\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.599124 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.600378 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.604852 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.605882 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-x89jc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.606190 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.606360 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.611149 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-6mrgc"] Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.614400 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.620875 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.624218 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.638734 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-6mrgc"] Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.642396 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-6mrgc\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.642443 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-config\") pod \"dnsmasq-dns-86db49b7ff-6mrgc\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.642464 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqqsj\" (UniqueName: \"kubernetes.io/projected/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-kube-api-access-fqqsj\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.642482 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.642605 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-6mrgc\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.642653 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk8kb\" (UniqueName: \"kubernetes.io/projected/1d66d04a-1fea-41a6-959b-e315ef09ac1b-kube-api-access-qk8kb\") pod \"dnsmasq-dns-86db49b7ff-6mrgc\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.642683 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t2k5\" (UniqueName: \"kubernetes.io/projected/dc44c0f6-5201-4b7a-85e9-db2a1766c237-kube-api-access-2t2k5\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.642701 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/dc44c0f6-5201-4b7a-85e9-db2a1766c237-ovs-rundir\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.642731 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/dc44c0f6-5201-4b7a-85e9-db2a1766c237-ovn-rundir\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.642905 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc44c0f6-5201-4b7a-85e9-db2a1766c237-config\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.642947 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.642966 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc44c0f6-5201-4b7a-85e9-db2a1766c237-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.643018 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.643038 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc44c0f6-5201-4b7a-85e9-db2a1766c237-combined-ca-bundle\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.643058 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-config\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.643087 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-scripts\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.643119 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-6mrgc\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.643160 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.643710 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/dc44c0f6-5201-4b7a-85e9-db2a1766c237-ovs-rundir\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.643793 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/dc44c0f6-5201-4b7a-85e9-db2a1766c237-ovn-rundir\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.644483 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc44c0f6-5201-4b7a-85e9-db2a1766c237-config\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.647927 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/dc44c0f6-5201-4b7a-85e9-db2a1766c237-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.651214 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc44c0f6-5201-4b7a-85e9-db2a1766c237-combined-ca-bundle\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.708303 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t2k5\" (UniqueName: \"kubernetes.io/projected/dc44c0f6-5201-4b7a-85e9-db2a1766c237-kube-api-access-2t2k5\") pod \"ovn-controller-metrics-hplrx\" (UID: \"dc44c0f6-5201-4b7a-85e9-db2a1766c237\") " pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.744549 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-6mrgc\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.744597 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk8kb\" (UniqueName: \"kubernetes.io/projected/1d66d04a-1fea-41a6-959b-e315ef09ac1b-kube-api-access-qk8kb\") pod \"dnsmasq-dns-86db49b7ff-6mrgc\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.744662 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.744693 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.744711 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-config\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.744730 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-scripts\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.744762 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-6mrgc\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.744792 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.744817 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-6mrgc\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.744840 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-config\") pod \"dnsmasq-dns-86db49b7ff-6mrgc\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.744858 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqqsj\" (UniqueName: \"kubernetes.io/projected/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-kube-api-access-fqqsj\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.744873 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.745499 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.745836 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-6mrgc\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.745867 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-6mrgc\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.745834 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-6mrgc\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.745836 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-config\") pod \"dnsmasq-dns-86db49b7ff-6mrgc\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.746371 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-config\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.752360 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-scripts\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.757359 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.760754 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.770535 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.780765 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk8kb\" (UniqueName: \"kubernetes.io/projected/1d66d04a-1fea-41a6-959b-e315ef09ac1b-kube-api-access-qk8kb\") pod \"dnsmasq-dns-86db49b7ff-6mrgc\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.781070 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-hplrx" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.782494 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqqsj\" (UniqueName: \"kubernetes.io/projected/ffe8e419-32a1-46e9-b2b3-a5f90c18dcde-kube-api-access-fqqsj\") pod \"ovn-northd-0\" (UID: \"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde\") " pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.933035 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 02 01:55:08 crc kubenswrapper[4884]: I1202 01:55:08.944130 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.163040 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.177473 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.266001 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-ovsdbserver-nb\") pod \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.266046 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-config\") pod \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.266105 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-dns-svc\") pod \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.266169 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dtx5\" (UniqueName: \"kubernetes.io/projected/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-kube-api-access-8dtx5\") pod \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\" (UID: \"bd50d32a-dcda-4403-aff4-0eabdbcaa6d8\") " Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.266612 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-config" (OuterVolumeSpecName: "config") pod "bd50d32a-dcda-4403-aff4-0eabdbcaa6d8" (UID: "bd50d32a-dcda-4403-aff4-0eabdbcaa6d8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.266877 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bd50d32a-dcda-4403-aff4-0eabdbcaa6d8" (UID: "bd50d32a-dcda-4403-aff4-0eabdbcaa6d8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.268771 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bd50d32a-dcda-4403-aff4-0eabdbcaa6d8" (UID: "bd50d32a-dcda-4403-aff4-0eabdbcaa6d8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.276710 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-kube-api-access-8dtx5" (OuterVolumeSpecName: "kube-api-access-8dtx5") pod "bd50d32a-dcda-4403-aff4-0eabdbcaa6d8" (UID: "bd50d32a-dcda-4403-aff4-0eabdbcaa6d8"). InnerVolumeSpecName "kube-api-access-8dtx5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.331458 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-hplrx"] Dec 02 01:55:09 crc kubenswrapper[4884]: W1202 01:55:09.345221 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc44c0f6_5201_4b7a_85e9_db2a1766c237.slice/crio-ebc14c708c2ac7451e6d5b369318a62add77aafd9819f7f498727423e400eec1 WatchSource:0}: Error finding container ebc14c708c2ac7451e6d5b369318a62add77aafd9819f7f498727423e400eec1: Status 404 returned error can't find the container with id ebc14c708c2ac7451e6d5b369318a62add77aafd9819f7f498727423e400eec1 Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.367438 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.367787 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.367846 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.367911 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dtx5\" (UniqueName: \"kubernetes.io/projected/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8-kube-api-access-8dtx5\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.499294 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.573564 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.573653 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.621853 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-6mrgc"] Dec 02 01:55:09 crc kubenswrapper[4884]: W1202 01:55:09.640087 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d66d04a_1fea_41a6_959b_e315ef09ac1b.slice/crio-6425d54bda196fc2149012a9c6d1ccfe2249866c7ac6e6adfad113ace0f5a03a WatchSource:0}: Error finding container 6425d54bda196fc2149012a9c6d1ccfe2249866c7ac6e6adfad113ace0f5a03a: Status 404 returned error can't find the container with id 6425d54bda196fc2149012a9c6d1ccfe2249866c7ac6e6adfad113ace0f5a03a Dec 02 01:55:09 crc kubenswrapper[4884]: I1202 01:55:09.889831 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 02 01:55:10 crc kubenswrapper[4884]: I1202 01:55:10.168502 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" event={"ID":"1d66d04a-1fea-41a6-959b-e315ef09ac1b","Type":"ContainerStarted","Data":"6425d54bda196fc2149012a9c6d1ccfe2249866c7ac6e6adfad113ace0f5a03a"} Dec 02 01:55:10 crc kubenswrapper[4884]: I1202 01:55:10.169930 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde","Type":"ContainerStarted","Data":"695adaef161a111ec916d6900f1d732a5d1eb1a6c635ddfec00dc6bcce1b130f"} Dec 02 01:55:10 crc kubenswrapper[4884]: I1202 01:55:10.170943 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-hplrx" event={"ID":"dc44c0f6-5201-4b7a-85e9-db2a1766c237","Type":"ContainerStarted","Data":"ebc14c708c2ac7451e6d5b369318a62add77aafd9819f7f498727423e400eec1"} Dec 02 01:55:10 crc kubenswrapper[4884]: I1202 01:55:10.171023 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-kvm9n" Dec 02 01:55:10 crc kubenswrapper[4884]: I1202 01:55:10.211118 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kvm9n"] Dec 02 01:55:10 crc kubenswrapper[4884]: I1202 01:55:10.220030 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-kvm9n"] Dec 02 01:55:11 crc kubenswrapper[4884]: I1202 01:55:11.021554 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 02 01:55:11 crc kubenswrapper[4884]: I1202 01:55:11.096197 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="63a48d7f-b557-4f21-93d6-8a18e96eea16" containerName="galera" probeResult="failure" output=< Dec 02 01:55:11 crc kubenswrapper[4884]: wsrep_local_state_comment (Joined) differs from Synced Dec 02 01:55:11 crc kubenswrapper[4884]: > Dec 02 01:55:11 crc kubenswrapper[4884]: I1202 01:55:11.626347 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd50d32a-dcda-4403-aff4-0eabdbcaa6d8" path="/var/lib/kubelet/pods/bd50d32a-dcda-4403-aff4-0eabdbcaa6d8/volumes" Dec 02 01:55:11 crc kubenswrapper[4884]: I1202 01:55:11.855712 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-6mrgc"] Dec 02 01:55:11 crc kubenswrapper[4884]: I1202 01:55:11.885277 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-5kq68"] Dec 02 01:55:11 crc kubenswrapper[4884]: I1202 01:55:11.886487 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:11 crc kubenswrapper[4884]: I1202 01:55:11.899053 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-5kq68"] Dec 02 01:55:11 crc kubenswrapper[4884]: I1202 01:55:11.950598 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.032678 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-dns-svc\") pod \"dnsmasq-dns-698758b865-5kq68\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.032774 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-5kq68\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.032827 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-config\") pod \"dnsmasq-dns-698758b865-5kq68\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.032853 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-5kq68\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.032885 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55k9t\" (UniqueName: \"kubernetes.io/projected/ff7b03af-7163-47fa-9397-ce25148d27c5-kube-api-access-55k9t\") pod \"dnsmasq-dns-698758b865-5kq68\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.133938 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-config\") pod \"dnsmasq-dns-698758b865-5kq68\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.133991 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-5kq68\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.134054 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55k9t\" (UniqueName: \"kubernetes.io/projected/ff7b03af-7163-47fa-9397-ce25148d27c5-kube-api-access-55k9t\") pod \"dnsmasq-dns-698758b865-5kq68\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.134998 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-config\") pod \"dnsmasq-dns-698758b865-5kq68\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.135067 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-5kq68\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.135637 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-dns-svc\") pod \"dnsmasq-dns-698758b865-5kq68\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.136364 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-dns-svc\") pod \"dnsmasq-dns-698758b865-5kq68\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.136509 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-5kq68\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.137237 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-5kq68\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.160689 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55k9t\" (UniqueName: \"kubernetes.io/projected/ff7b03af-7163-47fa-9397-ce25148d27c5-kube-api-access-55k9t\") pod \"dnsmasq-dns-698758b865-5kq68\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.187319 4884 generic.go:334] "Generic (PLEG): container finished" podID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerID="f632bbc70feae9bcab6e61183a7487126f6504963a2bb6247f36891b3ef48634" exitCode=0 Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.187360 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c8bb2562-5d06-44d1-bbd0-8848d92f18d2","Type":"ContainerDied","Data":"f632bbc70feae9bcab6e61183a7487126f6504963a2bb6247f36891b3ef48634"} Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.200219 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:12 crc kubenswrapper[4884]: I1202 01:55:12.702138 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-5kq68"] Dec 02 01:55:12 crc kubenswrapper[4884]: W1202 01:55:12.703903 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff7b03af_7163_47fa_9397_ce25148d27c5.slice/crio-50515de5c8fd063aa9355aeab5d9f9cbabe5330ef6e86b029d20da879387821e WatchSource:0}: Error finding container 50515de5c8fd063aa9355aeab5d9f9cbabe5330ef6e86b029d20da879387821e: Status 404 returned error can't find the container with id 50515de5c8fd063aa9355aeab5d9f9cbabe5330ef6e86b029d20da879387821e Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.001959 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.007865 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.014071 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.014448 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.014670 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.014942 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-nrf6z" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.038767 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.154961 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-cache\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.155041 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.155066 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-lock\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.155351 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.155505 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8wp8\" (UniqueName: \"kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-kube-api-access-f8wp8\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.196014 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-5kq68" event={"ID":"ff7b03af-7163-47fa-9397-ce25148d27c5","Type":"ContainerStarted","Data":"50515de5c8fd063aa9355aeab5d9f9cbabe5330ef6e86b029d20da879387821e"} Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.257104 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8wp8\" (UniqueName: \"kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-kube-api-access-f8wp8\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.257187 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-cache\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.257275 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.257317 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-lock\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.257421 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: E1202 01:55:13.257440 4884 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 01:55:13 crc kubenswrapper[4884]: E1202 01:55:13.257476 4884 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 01:55:13 crc kubenswrapper[4884]: E1202 01:55:13.257539 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift podName:8dc2a1f5-a7a6-4ce5-a711-77455efe2224 nodeName:}" failed. No retries permitted until 2025-12-02 01:55:13.7575153 +0000 UTC m=+1130.433352194 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift") pod "swift-storage-0" (UID: "8dc2a1f5-a7a6-4ce5-a711-77455efe2224") : configmap "swift-ring-files" not found Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.257760 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.257839 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-cache\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.258123 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-lock\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.282413 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.283331 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8wp8\" (UniqueName: \"kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-kube-api-access-f8wp8\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.393203 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-l2j2h"] Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.394368 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.396896 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.397348 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.397561 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.405611 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-l2j2h"] Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.562353 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-combined-ca-bundle\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.562431 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/151b4ede-9694-42d6-8051-6309de26a4e4-etc-swift\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.562567 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/151b4ede-9694-42d6-8051-6309de26a4e4-ring-data-devices\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.562644 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-dispersionconf\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.562714 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/151b4ede-9694-42d6-8051-6309de26a4e4-scripts\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.562780 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lld2z\" (UniqueName: \"kubernetes.io/projected/151b4ede-9694-42d6-8051-6309de26a4e4-kube-api-access-lld2z\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.562879 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-swiftconf\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.664315 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/151b4ede-9694-42d6-8051-6309de26a4e4-scripts\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.664411 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lld2z\" (UniqueName: \"kubernetes.io/projected/151b4ede-9694-42d6-8051-6309de26a4e4-kube-api-access-lld2z\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.664559 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-swiftconf\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.664724 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/151b4ede-9694-42d6-8051-6309de26a4e4-etc-swift\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.664806 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-combined-ca-bundle\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.664905 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/151b4ede-9694-42d6-8051-6309de26a4e4-ring-data-devices\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.665032 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-dispersionconf\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.665147 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/151b4ede-9694-42d6-8051-6309de26a4e4-scripts\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.665373 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/151b4ede-9694-42d6-8051-6309de26a4e4-etc-swift\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.665683 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/151b4ede-9694-42d6-8051-6309de26a4e4-ring-data-devices\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.669907 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-combined-ca-bundle\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.677234 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-dispersionconf\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.677346 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-swiftconf\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.762325 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lld2z\" (UniqueName: \"kubernetes.io/projected/151b4ede-9694-42d6-8051-6309de26a4e4-kube-api-access-lld2z\") pod \"swift-ring-rebalance-l2j2h\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:13 crc kubenswrapper[4884]: I1202 01:55:13.771339 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:13 crc kubenswrapper[4884]: E1202 01:55:13.773222 4884 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 01:55:13 crc kubenswrapper[4884]: E1202 01:55:13.773655 4884 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 01:55:13 crc kubenswrapper[4884]: E1202 01:55:13.773724 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift podName:8dc2a1f5-a7a6-4ce5-a711-77455efe2224 nodeName:}" failed. No retries permitted until 2025-12-02 01:55:14.773702338 +0000 UTC m=+1131.449539222 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift") pod "swift-storage-0" (UID: "8dc2a1f5-a7a6-4ce5-a711-77455efe2224") : configmap "swift-ring-files" not found Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.014106 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.212488 4884 generic.go:334] "Generic (PLEG): container finished" podID="1d66d04a-1fea-41a6-959b-e315ef09ac1b" containerID="8cb10709cb5573eb98e676be1c5f8853bd703d44196527ceef8926795b0660d2" exitCode=0 Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.212600 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" event={"ID":"1d66d04a-1fea-41a6-959b-e315ef09ac1b","Type":"ContainerDied","Data":"8cb10709cb5573eb98e676be1c5f8853bd703d44196527ceef8926795b0660d2"} Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.223892 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-hplrx" event={"ID":"dc44c0f6-5201-4b7a-85e9-db2a1766c237","Type":"ContainerStarted","Data":"25695b0101ff892cdd58345391e1e7210dda048845303463c366fa81014473ed"} Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.268908 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-hplrx" podStartSLOduration=6.268885858 podStartE2EDuration="6.268885858s" podCreationTimestamp="2025-12-02 01:55:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:55:14.260985699 +0000 UTC m=+1130.936822603" watchObservedRunningTime="2025-12-02 01:55:14.268885858 +0000 UTC m=+1130.944722742" Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.510182 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-l2j2h"] Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.593715 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.692171 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-config\") pod \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.692237 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-ovsdbserver-nb\") pod \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.692336 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-ovsdbserver-sb\") pod \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.692404 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qk8kb\" (UniqueName: \"kubernetes.io/projected/1d66d04a-1fea-41a6-959b-e315ef09ac1b-kube-api-access-qk8kb\") pod \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.692444 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-dns-svc\") pod \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\" (UID: \"1d66d04a-1fea-41a6-959b-e315ef09ac1b\") " Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.699241 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d66d04a-1fea-41a6-959b-e315ef09ac1b-kube-api-access-qk8kb" (OuterVolumeSpecName: "kube-api-access-qk8kb") pod "1d66d04a-1fea-41a6-959b-e315ef09ac1b" (UID: "1d66d04a-1fea-41a6-959b-e315ef09ac1b"). InnerVolumeSpecName "kube-api-access-qk8kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.715204 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-config" (OuterVolumeSpecName: "config") pod "1d66d04a-1fea-41a6-959b-e315ef09ac1b" (UID: "1d66d04a-1fea-41a6-959b-e315ef09ac1b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.719255 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1d66d04a-1fea-41a6-959b-e315ef09ac1b" (UID: "1d66d04a-1fea-41a6-959b-e315ef09ac1b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.722527 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1d66d04a-1fea-41a6-959b-e315ef09ac1b" (UID: "1d66d04a-1fea-41a6-959b-e315ef09ac1b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.727223 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1d66d04a-1fea-41a6-959b-e315ef09ac1b" (UID: "1d66d04a-1fea-41a6-959b-e315ef09ac1b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.796309 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.796478 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.796532 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qk8kb\" (UniqueName: \"kubernetes.io/projected/1d66d04a-1fea-41a6-959b-e315ef09ac1b-kube-api-access-qk8kb\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.796563 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.796581 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:14 crc kubenswrapper[4884]: I1202 01:55:14.796599 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d66d04a-1fea-41a6-959b-e315ef09ac1b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:14 crc kubenswrapper[4884]: E1202 01:55:14.796530 4884 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 01:55:14 crc kubenswrapper[4884]: E1202 01:55:14.796675 4884 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 01:55:14 crc kubenswrapper[4884]: E1202 01:55:14.796757 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift podName:8dc2a1f5-a7a6-4ce5-a711-77455efe2224 nodeName:}" failed. No retries permitted until 2025-12-02 01:55:16.796720889 +0000 UTC m=+1133.472557783 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift") pod "swift-storage-0" (UID: "8dc2a1f5-a7a6-4ce5-a711-77455efe2224") : configmap "swift-ring-files" not found Dec 02 01:55:15 crc kubenswrapper[4884]: I1202 01:55:15.233139 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-l2j2h" event={"ID":"151b4ede-9694-42d6-8051-6309de26a4e4","Type":"ContainerStarted","Data":"2dbe7c38a965077060c06ab05c2b00a16fafab5808d9b92887b2cd9f722c8b7b"} Dec 02 01:55:15 crc kubenswrapper[4884]: I1202 01:55:15.234619 4884 generic.go:334] "Generic (PLEG): container finished" podID="ff7b03af-7163-47fa-9397-ce25148d27c5" containerID="4124739ea4376e86d74c8f412d23ddbc69be6289c95f822b19b0763e879533b2" exitCode=0 Dec 02 01:55:15 crc kubenswrapper[4884]: I1202 01:55:15.234671 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-5kq68" event={"ID":"ff7b03af-7163-47fa-9397-ce25148d27c5","Type":"ContainerDied","Data":"4124739ea4376e86d74c8f412d23ddbc69be6289c95f822b19b0763e879533b2"} Dec 02 01:55:15 crc kubenswrapper[4884]: I1202 01:55:15.236451 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" event={"ID":"1d66d04a-1fea-41a6-959b-e315ef09ac1b","Type":"ContainerDied","Data":"6425d54bda196fc2149012a9c6d1ccfe2249866c7ac6e6adfad113ace0f5a03a"} Dec 02 01:55:15 crc kubenswrapper[4884]: I1202 01:55:15.236477 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-6mrgc" Dec 02 01:55:15 crc kubenswrapper[4884]: I1202 01:55:15.236507 4884 scope.go:117] "RemoveContainer" containerID="8cb10709cb5573eb98e676be1c5f8853bd703d44196527ceef8926795b0660d2" Dec 02 01:55:15 crc kubenswrapper[4884]: I1202 01:55:15.294245 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-6mrgc"] Dec 02 01:55:15 crc kubenswrapper[4884]: I1202 01:55:15.300388 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-6mrgc"] Dec 02 01:55:15 crc kubenswrapper[4884]: I1202 01:55:15.628982 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d66d04a-1fea-41a6-959b-e315ef09ac1b" path="/var/lib/kubelet/pods/1d66d04a-1fea-41a6-959b-e315ef09ac1b/volumes" Dec 02 01:55:16 crc kubenswrapper[4884]: I1202 01:55:16.247629 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde","Type":"ContainerStarted","Data":"abb64b2f7cde7c15130bcd6813f44ec83789a12cede6645363f905291afbaf36"} Dec 02 01:55:16 crc kubenswrapper[4884]: I1202 01:55:16.247680 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"ffe8e419-32a1-46e9-b2b3-a5f90c18dcde","Type":"ContainerStarted","Data":"86e3571b727709c64ee5a2031c62452cbcec8095d5fa510c6f21d8e205f988ea"} Dec 02 01:55:16 crc kubenswrapper[4884]: I1202 01:55:16.247778 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 02 01:55:16 crc kubenswrapper[4884]: I1202 01:55:16.251167 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-5kq68" event={"ID":"ff7b03af-7163-47fa-9397-ce25148d27c5","Type":"ContainerStarted","Data":"1be5bbc237238ae274ed7f2a27e79d37fddd40ccdf473e477fc749394117ff42"} Dec 02 01:55:16 crc kubenswrapper[4884]: I1202 01:55:16.251544 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:16 crc kubenswrapper[4884]: I1202 01:55:16.271157 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.317175326 podStartE2EDuration="8.271147228s" podCreationTimestamp="2025-12-02 01:55:08 +0000 UTC" firstStartedPulling="2025-12-02 01:55:09.504488227 +0000 UTC m=+1126.180325111" lastFinishedPulling="2025-12-02 01:55:15.458460129 +0000 UTC m=+1132.134297013" observedRunningTime="2025-12-02 01:55:16.266307236 +0000 UTC m=+1132.942144120" watchObservedRunningTime="2025-12-02 01:55:16.271147228 +0000 UTC m=+1132.946984112" Dec 02 01:55:16 crc kubenswrapper[4884]: I1202 01:55:16.294334 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-5kq68" podStartSLOduration=5.294315841 podStartE2EDuration="5.294315841s" podCreationTimestamp="2025-12-02 01:55:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:55:16.288699549 +0000 UTC m=+1132.964536433" watchObservedRunningTime="2025-12-02 01:55:16.294315841 +0000 UTC m=+1132.970152735" Dec 02 01:55:16 crc kubenswrapper[4884]: I1202 01:55:16.835364 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:16 crc kubenswrapper[4884]: E1202 01:55:16.835637 4884 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 01:55:16 crc kubenswrapper[4884]: E1202 01:55:16.835660 4884 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 01:55:16 crc kubenswrapper[4884]: E1202 01:55:16.835732 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift podName:8dc2a1f5-a7a6-4ce5-a711-77455efe2224 nodeName:}" failed. No retries permitted until 2025-12-02 01:55:20.835714603 +0000 UTC m=+1137.511551487 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift") pod "swift-storage-0" (UID: "8dc2a1f5-a7a6-4ce5-a711-77455efe2224") : configmap "swift-ring-files" not found Dec 02 01:55:17 crc kubenswrapper[4884]: I1202 01:55:17.700895 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 02 01:55:17 crc kubenswrapper[4884]: I1202 01:55:17.784049 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 02 01:55:18 crc kubenswrapper[4884]: I1202 01:55:18.095602 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.564870 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8791-account-create-update-j27h4"] Dec 02 01:55:19 crc kubenswrapper[4884]: E1202 01:55:19.565261 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d66d04a-1fea-41a6-959b-e315ef09ac1b" containerName="init" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.565277 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d66d04a-1fea-41a6-959b-e315ef09ac1b" containerName="init" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.565508 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d66d04a-1fea-41a6-959b-e315ef09ac1b" containerName="init" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.576344 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8791-account-create-update-j27h4" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.579721 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.594021 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8791-account-create-update-j27h4"] Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.604271 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-9g2b5"] Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.605466 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9g2b5" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.612160 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12142209-d074-426f-8e5d-d9e4acae21a0-operator-scripts\") pod \"keystone-8791-account-create-update-j27h4\" (UID: \"12142209-d074-426f-8e5d-d9e4acae21a0\") " pod="openstack/keystone-8791-account-create-update-j27h4" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.612275 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcqd4\" (UniqueName: \"kubernetes.io/projected/12142209-d074-426f-8e5d-d9e4acae21a0-kube-api-access-xcqd4\") pod \"keystone-8791-account-create-update-j27h4\" (UID: \"12142209-d074-426f-8e5d-d9e4acae21a0\") " pod="openstack/keystone-8791-account-create-update-j27h4" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.636778 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9g2b5"] Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.713459 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c764da5f-6cfc-48aa-8a10-bf6bd97a226f-operator-scripts\") pod \"keystone-db-create-9g2b5\" (UID: \"c764da5f-6cfc-48aa-8a10-bf6bd97a226f\") " pod="openstack/keystone-db-create-9g2b5" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.713546 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcqd4\" (UniqueName: \"kubernetes.io/projected/12142209-d074-426f-8e5d-d9e4acae21a0-kube-api-access-xcqd4\") pod \"keystone-8791-account-create-update-j27h4\" (UID: \"12142209-d074-426f-8e5d-d9e4acae21a0\") " pod="openstack/keystone-8791-account-create-update-j27h4" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.713611 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xbc6\" (UniqueName: \"kubernetes.io/projected/c764da5f-6cfc-48aa-8a10-bf6bd97a226f-kube-api-access-5xbc6\") pod \"keystone-db-create-9g2b5\" (UID: \"c764da5f-6cfc-48aa-8a10-bf6bd97a226f\") " pod="openstack/keystone-db-create-9g2b5" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.713630 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12142209-d074-426f-8e5d-d9e4acae21a0-operator-scripts\") pod \"keystone-8791-account-create-update-j27h4\" (UID: \"12142209-d074-426f-8e5d-d9e4acae21a0\") " pod="openstack/keystone-8791-account-create-update-j27h4" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.714280 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12142209-d074-426f-8e5d-d9e4acae21a0-operator-scripts\") pod \"keystone-8791-account-create-update-j27h4\" (UID: \"12142209-d074-426f-8e5d-d9e4acae21a0\") " pod="openstack/keystone-8791-account-create-update-j27h4" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.733168 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcqd4\" (UniqueName: \"kubernetes.io/projected/12142209-d074-426f-8e5d-d9e4acae21a0-kube-api-access-xcqd4\") pod \"keystone-8791-account-create-update-j27h4\" (UID: \"12142209-d074-426f-8e5d-d9e4acae21a0\") " pod="openstack/keystone-8791-account-create-update-j27h4" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.783652 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-dvqcv"] Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.784902 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dvqcv" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.798420 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-dvqcv"] Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.814642 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4nxt\" (UniqueName: \"kubernetes.io/projected/7f9006f4-9196-4a39-8c4d-10c4344a3b1f-kube-api-access-p4nxt\") pod \"placement-db-create-dvqcv\" (UID: \"7f9006f4-9196-4a39-8c4d-10c4344a3b1f\") " pod="openstack/placement-db-create-dvqcv" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.814798 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xbc6\" (UniqueName: \"kubernetes.io/projected/c764da5f-6cfc-48aa-8a10-bf6bd97a226f-kube-api-access-5xbc6\") pod \"keystone-db-create-9g2b5\" (UID: \"c764da5f-6cfc-48aa-8a10-bf6bd97a226f\") " pod="openstack/keystone-db-create-9g2b5" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.815037 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c764da5f-6cfc-48aa-8a10-bf6bd97a226f-operator-scripts\") pod \"keystone-db-create-9g2b5\" (UID: \"c764da5f-6cfc-48aa-8a10-bf6bd97a226f\") " pod="openstack/keystone-db-create-9g2b5" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.815293 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f9006f4-9196-4a39-8c4d-10c4344a3b1f-operator-scripts\") pod \"placement-db-create-dvqcv\" (UID: \"7f9006f4-9196-4a39-8c4d-10c4344a3b1f\") " pod="openstack/placement-db-create-dvqcv" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.816735 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c764da5f-6cfc-48aa-8a10-bf6bd97a226f-operator-scripts\") pod \"keystone-db-create-9g2b5\" (UID: \"c764da5f-6cfc-48aa-8a10-bf6bd97a226f\") " pod="openstack/keystone-db-create-9g2b5" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.836424 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xbc6\" (UniqueName: \"kubernetes.io/projected/c764da5f-6cfc-48aa-8a10-bf6bd97a226f-kube-api-access-5xbc6\") pod \"keystone-db-create-9g2b5\" (UID: \"c764da5f-6cfc-48aa-8a10-bf6bd97a226f\") " pod="openstack/keystone-db-create-9g2b5" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.855853 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-4d9f-account-create-update-bc8h8"] Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.857823 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4d9f-account-create-update-bc8h8" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.860120 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.866432 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4d9f-account-create-update-bc8h8"] Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.907802 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8791-account-create-update-j27h4" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.916182 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f9006f4-9196-4a39-8c4d-10c4344a3b1f-operator-scripts\") pod \"placement-db-create-dvqcv\" (UID: \"7f9006f4-9196-4a39-8c4d-10c4344a3b1f\") " pod="openstack/placement-db-create-dvqcv" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.916242 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4nxt\" (UniqueName: \"kubernetes.io/projected/7f9006f4-9196-4a39-8c4d-10c4344a3b1f-kube-api-access-p4nxt\") pod \"placement-db-create-dvqcv\" (UID: \"7f9006f4-9196-4a39-8c4d-10c4344a3b1f\") " pod="openstack/placement-db-create-dvqcv" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.917083 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f9006f4-9196-4a39-8c4d-10c4344a3b1f-operator-scripts\") pod \"placement-db-create-dvqcv\" (UID: \"7f9006f4-9196-4a39-8c4d-10c4344a3b1f\") " pod="openstack/placement-db-create-dvqcv" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.923003 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9g2b5" Dec 02 01:55:19 crc kubenswrapper[4884]: I1202 01:55:19.931237 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4nxt\" (UniqueName: \"kubernetes.io/projected/7f9006f4-9196-4a39-8c4d-10c4344a3b1f-kube-api-access-p4nxt\") pod \"placement-db-create-dvqcv\" (UID: \"7f9006f4-9196-4a39-8c4d-10c4344a3b1f\") " pod="openstack/placement-db-create-dvqcv" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.007775 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-8g74n"] Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.008905 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8g74n" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.019596 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxmp9\" (UniqueName: \"kubernetes.io/projected/9922aaf4-6776-4e6c-ac62-82c43bef4697-kube-api-access-rxmp9\") pod \"glance-db-create-8g74n\" (UID: \"9922aaf4-6776-4e6c-ac62-82c43bef4697\") " pod="openstack/glance-db-create-8g74n" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.019673 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05-operator-scripts\") pod \"placement-4d9f-account-create-update-bc8h8\" (UID: \"e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05\") " pod="openstack/placement-4d9f-account-create-update-bc8h8" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.020057 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnmxl\" (UniqueName: \"kubernetes.io/projected/e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05-kube-api-access-tnmxl\") pod \"placement-4d9f-account-create-update-bc8h8\" (UID: \"e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05\") " pod="openstack/placement-4d9f-account-create-update-bc8h8" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.020106 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9922aaf4-6776-4e6c-ac62-82c43bef4697-operator-scripts\") pod \"glance-db-create-8g74n\" (UID: \"9922aaf4-6776-4e6c-ac62-82c43bef4697\") " pod="openstack/glance-db-create-8g74n" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.027101 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8g74n"] Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.101890 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dvqcv" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.120928 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnmxl\" (UniqueName: \"kubernetes.io/projected/e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05-kube-api-access-tnmxl\") pod \"placement-4d9f-account-create-update-bc8h8\" (UID: \"e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05\") " pod="openstack/placement-4d9f-account-create-update-bc8h8" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.120979 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9922aaf4-6776-4e6c-ac62-82c43bef4697-operator-scripts\") pod \"glance-db-create-8g74n\" (UID: \"9922aaf4-6776-4e6c-ac62-82c43bef4697\") " pod="openstack/glance-db-create-8g74n" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.121062 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxmp9\" (UniqueName: \"kubernetes.io/projected/9922aaf4-6776-4e6c-ac62-82c43bef4697-kube-api-access-rxmp9\") pod \"glance-db-create-8g74n\" (UID: \"9922aaf4-6776-4e6c-ac62-82c43bef4697\") " pod="openstack/glance-db-create-8g74n" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.121110 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05-operator-scripts\") pod \"placement-4d9f-account-create-update-bc8h8\" (UID: \"e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05\") " pod="openstack/placement-4d9f-account-create-update-bc8h8" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.121834 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05-operator-scripts\") pod \"placement-4d9f-account-create-update-bc8h8\" (UID: \"e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05\") " pod="openstack/placement-4d9f-account-create-update-bc8h8" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.121851 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9922aaf4-6776-4e6c-ac62-82c43bef4697-operator-scripts\") pod \"glance-db-create-8g74n\" (UID: \"9922aaf4-6776-4e6c-ac62-82c43bef4697\") " pod="openstack/glance-db-create-8g74n" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.140889 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxmp9\" (UniqueName: \"kubernetes.io/projected/9922aaf4-6776-4e6c-ac62-82c43bef4697-kube-api-access-rxmp9\") pod \"glance-db-create-8g74n\" (UID: \"9922aaf4-6776-4e6c-ac62-82c43bef4697\") " pod="openstack/glance-db-create-8g74n" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.158457 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnmxl\" (UniqueName: \"kubernetes.io/projected/e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05-kube-api-access-tnmxl\") pod \"placement-4d9f-account-create-update-bc8h8\" (UID: \"e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05\") " pod="openstack/placement-4d9f-account-create-update-bc8h8" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.163602 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-4c10-account-create-update-kmph9"] Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.164729 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4c10-account-create-update-kmph9" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.167084 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.183804 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-4c10-account-create-update-kmph9"] Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.195925 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4d9f-account-create-update-bc8h8" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.222768 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htr7l\" (UniqueName: \"kubernetes.io/projected/1c11795b-a2e2-4906-b3fd-3048ec29a764-kube-api-access-htr7l\") pod \"glance-4c10-account-create-update-kmph9\" (UID: \"1c11795b-a2e2-4906-b3fd-3048ec29a764\") " pod="openstack/glance-4c10-account-create-update-kmph9" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.222944 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c11795b-a2e2-4906-b3fd-3048ec29a764-operator-scripts\") pod \"glance-4c10-account-create-update-kmph9\" (UID: \"1c11795b-a2e2-4906-b3fd-3048ec29a764\") " pod="openstack/glance-4c10-account-create-update-kmph9" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.324376 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c11795b-a2e2-4906-b3fd-3048ec29a764-operator-scripts\") pod \"glance-4c10-account-create-update-kmph9\" (UID: \"1c11795b-a2e2-4906-b3fd-3048ec29a764\") " pod="openstack/glance-4c10-account-create-update-kmph9" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.324465 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htr7l\" (UniqueName: \"kubernetes.io/projected/1c11795b-a2e2-4906-b3fd-3048ec29a764-kube-api-access-htr7l\") pod \"glance-4c10-account-create-update-kmph9\" (UID: \"1c11795b-a2e2-4906-b3fd-3048ec29a764\") " pod="openstack/glance-4c10-account-create-update-kmph9" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.325394 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c11795b-a2e2-4906-b3fd-3048ec29a764-operator-scripts\") pod \"glance-4c10-account-create-update-kmph9\" (UID: \"1c11795b-a2e2-4906-b3fd-3048ec29a764\") " pod="openstack/glance-4c10-account-create-update-kmph9" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.334257 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8g74n" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.360345 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htr7l\" (UniqueName: \"kubernetes.io/projected/1c11795b-a2e2-4906-b3fd-3048ec29a764-kube-api-access-htr7l\") pod \"glance-4c10-account-create-update-kmph9\" (UID: \"1c11795b-a2e2-4906-b3fd-3048ec29a764\") " pod="openstack/glance-4c10-account-create-update-kmph9" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.552960 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4c10-account-create-update-kmph9" Dec 02 01:55:20 crc kubenswrapper[4884]: I1202 01:55:20.935589 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:20 crc kubenswrapper[4884]: E1202 01:55:20.935822 4884 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 01:55:20 crc kubenswrapper[4884]: E1202 01:55:20.935863 4884 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 01:55:20 crc kubenswrapper[4884]: E1202 01:55:20.935956 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift podName:8dc2a1f5-a7a6-4ce5-a711-77455efe2224 nodeName:}" failed. No retries permitted until 2025-12-02 01:55:28.935928222 +0000 UTC m=+1145.611765136 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift") pod "swift-storage-0" (UID: "8dc2a1f5-a7a6-4ce5-a711-77455efe2224") : configmap "swift-ring-files" not found Dec 02 01:55:21 crc kubenswrapper[4884]: I1202 01:55:21.856197 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-create-484hp"] Dec 02 01:55:21 crc kubenswrapper[4884]: I1202 01:55:21.858050 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-484hp" Dec 02 01:55:21 crc kubenswrapper[4884]: I1202 01:55:21.863849 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-484hp"] Dec 02 01:55:21 crc kubenswrapper[4884]: I1202 01:55:21.952481 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-355c-account-create-update-d5b2t"] Dec 02 01:55:21 crc kubenswrapper[4884]: I1202 01:55:21.954319 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-355c-account-create-update-d5b2t" Dec 02 01:55:21 crc kubenswrapper[4884]: I1202 01:55:21.958323 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-db-secret" Dec 02 01:55:21 crc kubenswrapper[4884]: I1202 01:55:21.975575 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-355c-account-create-update-d5b2t"] Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.052972 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbf8n\" (UniqueName: \"kubernetes.io/projected/a0bb5aee-14d4-4023-bb79-fdd1824aae0e-kube-api-access-cbf8n\") pod \"watcher-db-create-484hp\" (UID: \"a0bb5aee-14d4-4023-bb79-fdd1824aae0e\") " pod="openstack/watcher-db-create-484hp" Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.053162 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0bb5aee-14d4-4023-bb79-fdd1824aae0e-operator-scripts\") pod \"watcher-db-create-484hp\" (UID: \"a0bb5aee-14d4-4023-bb79-fdd1824aae0e\") " pod="openstack/watcher-db-create-484hp" Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.155054 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbf8n\" (UniqueName: \"kubernetes.io/projected/a0bb5aee-14d4-4023-bb79-fdd1824aae0e-kube-api-access-cbf8n\") pod \"watcher-db-create-484hp\" (UID: \"a0bb5aee-14d4-4023-bb79-fdd1824aae0e\") " pod="openstack/watcher-db-create-484hp" Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.155348 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0bb5aee-14d4-4023-bb79-fdd1824aae0e-operator-scripts\") pod \"watcher-db-create-484hp\" (UID: \"a0bb5aee-14d4-4023-bb79-fdd1824aae0e\") " pod="openstack/watcher-db-create-484hp" Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.155394 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33e4ef6f-578b-4149-8d23-fcf1b3f4a137-operator-scripts\") pod \"watcher-355c-account-create-update-d5b2t\" (UID: \"33e4ef6f-578b-4149-8d23-fcf1b3f4a137\") " pod="openstack/watcher-355c-account-create-update-d5b2t" Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.155464 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stbrp\" (UniqueName: \"kubernetes.io/projected/33e4ef6f-578b-4149-8d23-fcf1b3f4a137-kube-api-access-stbrp\") pod \"watcher-355c-account-create-update-d5b2t\" (UID: \"33e4ef6f-578b-4149-8d23-fcf1b3f4a137\") " pod="openstack/watcher-355c-account-create-update-d5b2t" Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.156980 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0bb5aee-14d4-4023-bb79-fdd1824aae0e-operator-scripts\") pod \"watcher-db-create-484hp\" (UID: \"a0bb5aee-14d4-4023-bb79-fdd1824aae0e\") " pod="openstack/watcher-db-create-484hp" Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.173441 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbf8n\" (UniqueName: \"kubernetes.io/projected/a0bb5aee-14d4-4023-bb79-fdd1824aae0e-kube-api-access-cbf8n\") pod \"watcher-db-create-484hp\" (UID: \"a0bb5aee-14d4-4023-bb79-fdd1824aae0e\") " pod="openstack/watcher-db-create-484hp" Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.180496 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-484hp" Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.202880 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.261939 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33e4ef6f-578b-4149-8d23-fcf1b3f4a137-operator-scripts\") pod \"watcher-355c-account-create-update-d5b2t\" (UID: \"33e4ef6f-578b-4149-8d23-fcf1b3f4a137\") " pod="openstack/watcher-355c-account-create-update-d5b2t" Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.262033 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stbrp\" (UniqueName: \"kubernetes.io/projected/33e4ef6f-578b-4149-8d23-fcf1b3f4a137-kube-api-access-stbrp\") pod \"watcher-355c-account-create-update-d5b2t\" (UID: \"33e4ef6f-578b-4149-8d23-fcf1b3f4a137\") " pod="openstack/watcher-355c-account-create-update-d5b2t" Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.263029 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33e4ef6f-578b-4149-8d23-fcf1b3f4a137-operator-scripts\") pod \"watcher-355c-account-create-update-d5b2t\" (UID: \"33e4ef6f-578b-4149-8d23-fcf1b3f4a137\") " pod="openstack/watcher-355c-account-create-update-d5b2t" Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.267452 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tf8jx"] Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.267733 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" podUID="01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d" containerName="dnsmasq-dns" containerID="cri-o://256f1c2d98576266fed9b5f4578d9424e93076aeaf6f42686017f7323a6dfa1f" gracePeriod=10 Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.298905 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stbrp\" (UniqueName: \"kubernetes.io/projected/33e4ef6f-578b-4149-8d23-fcf1b3f4a137-kube-api-access-stbrp\") pod \"watcher-355c-account-create-update-d5b2t\" (UID: \"33e4ef6f-578b-4149-8d23-fcf1b3f4a137\") " pod="openstack/watcher-355c-account-create-update-d5b2t" Dec 02 01:55:22 crc kubenswrapper[4884]: I1202 01:55:22.572772 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-355c-account-create-update-d5b2t" Dec 02 01:55:23 crc kubenswrapper[4884]: I1202 01:55:23.322725 4884 generic.go:334] "Generic (PLEG): container finished" podID="01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d" containerID="256f1c2d98576266fed9b5f4578d9424e93076aeaf6f42686017f7323a6dfa1f" exitCode=0 Dec 02 01:55:23 crc kubenswrapper[4884]: I1202 01:55:23.322842 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" event={"ID":"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d","Type":"ContainerDied","Data":"256f1c2d98576266fed9b5f4578d9424e93076aeaf6f42686017f7323a6dfa1f"} Dec 02 01:55:23 crc kubenswrapper[4884]: I1202 01:55:23.818576 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" Dec 02 01:55:23 crc kubenswrapper[4884]: I1202 01:55:23.908166 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-config\") pod \"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d\" (UID: \"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d\") " Dec 02 01:55:23 crc kubenswrapper[4884]: I1202 01:55:23.908209 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-dns-svc\") pod \"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d\" (UID: \"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d\") " Dec 02 01:55:23 crc kubenswrapper[4884]: I1202 01:55:23.908233 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd9xn\" (UniqueName: \"kubernetes.io/projected/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-kube-api-access-bd9xn\") pod \"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d\" (UID: \"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d\") " Dec 02 01:55:23 crc kubenswrapper[4884]: I1202 01:55:23.927732 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-kube-api-access-bd9xn" (OuterVolumeSpecName: "kube-api-access-bd9xn") pod "01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d" (UID: "01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d"). InnerVolumeSpecName "kube-api-access-bd9xn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:23 crc kubenswrapper[4884]: I1202 01:55:23.991901 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-config" (OuterVolumeSpecName: "config") pod "01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d" (UID: "01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:23 crc kubenswrapper[4884]: I1202 01:55:23.994930 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d" (UID: "01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.018770 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.018802 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.018814 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bd9xn\" (UniqueName: \"kubernetes.io/projected/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d-kube-api-access-bd9xn\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.049519 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-dvqcv"] Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.272834 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9g2b5"] Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.283897 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-4d9f-account-create-update-bc8h8"] Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.293614 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8g74n"] Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.302898 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-4c10-account-create-update-kmph9"] Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.311287 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-355c-account-create-update-d5b2t"] Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.320657 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.334765 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.354456 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-db-secret" Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.378931 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-355c-account-create-update-d5b2t" event={"ID":"33e4ef6f-578b-4149-8d23-fcf1b3f4a137","Type":"ContainerStarted","Data":"f77a0300f4f70f744d9a8a4de7aedab343e6199aa089b6d950c0932950d214d0"} Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.384562 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" event={"ID":"01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d","Type":"ContainerDied","Data":"9acc15b2d1dae6995338120a46845009c9a5df6327139342c5349e4a616bed31"} Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.384628 4884 scope.go:117] "RemoveContainer" containerID="256f1c2d98576266fed9b5f4578d9424e93076aeaf6f42686017f7323a6dfa1f" Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.384799 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-tf8jx" Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.394650 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8g74n" event={"ID":"9922aaf4-6776-4e6c-ac62-82c43bef4697","Type":"ContainerStarted","Data":"1aac1696eb44fdb0163d26ae614d375a1fc4e853072fe435ca0fcff3fdce8a75"} Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.397875 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4c10-account-create-update-kmph9" event={"ID":"1c11795b-a2e2-4906-b3fd-3048ec29a764","Type":"ContainerStarted","Data":"783e13ee39326019cfff3e54328d32a27ac69ad161c4be419d7b0b418c8e341f"} Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.405553 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9g2b5" event={"ID":"c764da5f-6cfc-48aa-8a10-bf6bd97a226f","Type":"ContainerStarted","Data":"d422ba352d956106286ca781e5a02094801b4bdbd7315ec4f77ccf8b2f0dc340"} Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.407386 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4d9f-account-create-update-bc8h8" event={"ID":"e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05","Type":"ContainerStarted","Data":"d3b5475f60b0069b3845972199410debe6b71383b9efd2a9e435647268ebfdb4"} Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.409722 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-dvqcv" event={"ID":"7f9006f4-9196-4a39-8c4d-10c4344a3b1f","Type":"ContainerStarted","Data":"7a4aa8b576a5faba0676a0107a1e59151a893f698b06b25a73e03328a444cc50"} Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.409779 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-dvqcv" event={"ID":"7f9006f4-9196-4a39-8c4d-10c4344a3b1f","Type":"ContainerStarted","Data":"2f291e130e58a5beabfcb89379832ccb0d526cf2c51c7236e190d86be6e226d6"} Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.412897 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-l2j2h" event={"ID":"151b4ede-9694-42d6-8051-6309de26a4e4","Type":"ContainerStarted","Data":"fdc7b71ec90bc36daeab689156c99d0836cfd20761769249ac79beee976a010f"} Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.416601 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c8bb2562-5d06-44d1-bbd0-8848d92f18d2","Type":"ContainerStarted","Data":"05e715ded10f8d73e952217f675ad561ba62a82b52a6306b6dd3b7b674d8329e"} Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.428632 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-dvqcv" podStartSLOduration=5.4286152229999995 podStartE2EDuration="5.428615223s" podCreationTimestamp="2025-12-02 01:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:55:24.428302986 +0000 UTC m=+1141.104139880" watchObservedRunningTime="2025-12-02 01:55:24.428615223 +0000 UTC m=+1141.104452107" Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.493588 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-l2j2h" podStartSLOduration=2.486140047 podStartE2EDuration="11.493571638s" podCreationTimestamp="2025-12-02 01:55:13 +0000 UTC" firstStartedPulling="2025-12-02 01:55:14.551058358 +0000 UTC m=+1131.226895242" lastFinishedPulling="2025-12-02 01:55:23.558489949 +0000 UTC m=+1140.234326833" observedRunningTime="2025-12-02 01:55:24.445900278 +0000 UTC m=+1141.121737182" watchObservedRunningTime="2025-12-02 01:55:24.493571638 +0000 UTC m=+1141.169408532" Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.498646 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8791-account-create-update-j27h4"] Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.519734 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-create-484hp"] Dec 02 01:55:24 crc kubenswrapper[4884]: W1202 01:55:24.550987 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0bb5aee_14d4_4023_bb79_fdd1824aae0e.slice/crio-01e33d7675021eba461367e1f50b83f3c60ba7664bcb505845c569a1c9dc6df3 WatchSource:0}: Error finding container 01e33d7675021eba461367e1f50b83f3c60ba7664bcb505845c569a1c9dc6df3: Status 404 returned error can't find the container with id 01e33d7675021eba461367e1f50b83f3c60ba7664bcb505845c569a1c9dc6df3 Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.554560 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.628862 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tf8jx"] Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.634573 4884 scope.go:117] "RemoveContainer" containerID="27ba7839a91e5ed64eb8095c1c7248d8a230bc2665f1b4ffeab8f98debcd9c8a" Dec 02 01:55:24 crc kubenswrapper[4884]: I1202 01:55:24.639867 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-tf8jx"] Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.428411 4884 generic.go:334] "Generic (PLEG): container finished" podID="33e4ef6f-578b-4149-8d23-fcf1b3f4a137" containerID="0009a1795b6b89769018f7bf4a7a02b1bb07a26d36f296e87ebb240ca8cc84bf" exitCode=0 Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.428524 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-355c-account-create-update-d5b2t" event={"ID":"33e4ef6f-578b-4149-8d23-fcf1b3f4a137","Type":"ContainerDied","Data":"0009a1795b6b89769018f7bf4a7a02b1bb07a26d36f296e87ebb240ca8cc84bf"} Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.448479 4884 generic.go:334] "Generic (PLEG): container finished" podID="a0bb5aee-14d4-4023-bb79-fdd1824aae0e" containerID="73dd6369a1ef29bdb1afed9fc8e62c11012312a273172268c8b4cf1b454abd4b" exitCode=0 Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.448736 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-484hp" event={"ID":"a0bb5aee-14d4-4023-bb79-fdd1824aae0e","Type":"ContainerDied","Data":"73dd6369a1ef29bdb1afed9fc8e62c11012312a273172268c8b4cf1b454abd4b"} Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.448791 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-484hp" event={"ID":"a0bb5aee-14d4-4023-bb79-fdd1824aae0e","Type":"ContainerStarted","Data":"01e33d7675021eba461367e1f50b83f3c60ba7664bcb505845c569a1c9dc6df3"} Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.451847 4884 generic.go:334] "Generic (PLEG): container finished" podID="1c11795b-a2e2-4906-b3fd-3048ec29a764" containerID="27ee94a09469808f8d08fa4e19c1a06b2a625b5badff4d05ceb225571b42d96e" exitCode=0 Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.451879 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4c10-account-create-update-kmph9" event={"ID":"1c11795b-a2e2-4906-b3fd-3048ec29a764","Type":"ContainerDied","Data":"27ee94a09469808f8d08fa4e19c1a06b2a625b5badff4d05ceb225571b42d96e"} Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.453810 4884 generic.go:334] "Generic (PLEG): container finished" podID="7f9006f4-9196-4a39-8c4d-10c4344a3b1f" containerID="7a4aa8b576a5faba0676a0107a1e59151a893f698b06b25a73e03328a444cc50" exitCode=0 Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.453918 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-dvqcv" event={"ID":"7f9006f4-9196-4a39-8c4d-10c4344a3b1f","Type":"ContainerDied","Data":"7a4aa8b576a5faba0676a0107a1e59151a893f698b06b25a73e03328a444cc50"} Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.462670 4884 generic.go:334] "Generic (PLEG): container finished" podID="e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05" containerID="d483bb9d56e217aa939f6edd2769b9496f0c550ec29df470b9b4461c520bfc7b" exitCode=0 Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.462776 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4d9f-account-create-update-bc8h8" event={"ID":"e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05","Type":"ContainerDied","Data":"d483bb9d56e217aa939f6edd2769b9496f0c550ec29df470b9b4461c520bfc7b"} Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.464951 4884 generic.go:334] "Generic (PLEG): container finished" podID="9922aaf4-6776-4e6c-ac62-82c43bef4697" containerID="99b6d45eb80119a04d35704de720f35dc87bad121a63e34001944bb5d691f4a0" exitCode=0 Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.465031 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8g74n" event={"ID":"9922aaf4-6776-4e6c-ac62-82c43bef4697","Type":"ContainerDied","Data":"99b6d45eb80119a04d35704de720f35dc87bad121a63e34001944bb5d691f4a0"} Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.467243 4884 generic.go:334] "Generic (PLEG): container finished" podID="c764da5f-6cfc-48aa-8a10-bf6bd97a226f" containerID="eea4c4a5937e70803834d77ed39c5a35944c2c5183e5e7b9c57c75d5072f961f" exitCode=0 Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.467321 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9g2b5" event={"ID":"c764da5f-6cfc-48aa-8a10-bf6bd97a226f","Type":"ContainerDied","Data":"eea4c4a5937e70803834d77ed39c5a35944c2c5183e5e7b9c57c75d5072f961f"} Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.469340 4884 generic.go:334] "Generic (PLEG): container finished" podID="12142209-d074-426f-8e5d-d9e4acae21a0" containerID="253198cc432448f0046c67fd4e99c31cf153588dc26bfb91548aa7651d20489c" exitCode=0 Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.469456 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8791-account-create-update-j27h4" event={"ID":"12142209-d074-426f-8e5d-d9e4acae21a0","Type":"ContainerDied","Data":"253198cc432448f0046c67fd4e99c31cf153588dc26bfb91548aa7651d20489c"} Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.469522 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8791-account-create-update-j27h4" event={"ID":"12142209-d074-426f-8e5d-d9e4acae21a0","Type":"ContainerStarted","Data":"0615344d80c71e58ce944d2b37ef5cf4722267f310ee747b8795ee904a285e02"} Dec 02 01:55:25 crc kubenswrapper[4884]: I1202 01:55:25.629024 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d" path="/var/lib/kubelet/pods/01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d/volumes" Dec 02 01:55:26 crc kubenswrapper[4884]: I1202 01:55:26.486012 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c8bb2562-5d06-44d1-bbd0-8848d92f18d2","Type":"ContainerStarted","Data":"38a9fea55138df4195eab6c884265c3d5ff494f21607ba70544cff03fee07d66"} Dec 02 01:55:26 crc kubenswrapper[4884]: I1202 01:55:26.969260 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8g74n" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.082696 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9922aaf4-6776-4e6c-ac62-82c43bef4697-operator-scripts\") pod \"9922aaf4-6776-4e6c-ac62-82c43bef4697\" (UID: \"9922aaf4-6776-4e6c-ac62-82c43bef4697\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.082753 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxmp9\" (UniqueName: \"kubernetes.io/projected/9922aaf4-6776-4e6c-ac62-82c43bef4697-kube-api-access-rxmp9\") pod \"9922aaf4-6776-4e6c-ac62-82c43bef4697\" (UID: \"9922aaf4-6776-4e6c-ac62-82c43bef4697\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.091980 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9922aaf4-6776-4e6c-ac62-82c43bef4697-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9922aaf4-6776-4e6c-ac62-82c43bef4697" (UID: "9922aaf4-6776-4e6c-ac62-82c43bef4697"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.104086 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9922aaf4-6776-4e6c-ac62-82c43bef4697-kube-api-access-rxmp9" (OuterVolumeSpecName: "kube-api-access-rxmp9") pod "9922aaf4-6776-4e6c-ac62-82c43bef4697" (UID: "9922aaf4-6776-4e6c-ac62-82c43bef4697"). InnerVolumeSpecName "kube-api-access-rxmp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.142579 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9g2b5" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.154789 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-355c-account-create-update-d5b2t" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.169067 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4d9f-account-create-update-bc8h8" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.173451 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-484hp" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.185332 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stbrp\" (UniqueName: \"kubernetes.io/projected/33e4ef6f-578b-4149-8d23-fcf1b3f4a137-kube-api-access-stbrp\") pod \"33e4ef6f-578b-4149-8d23-fcf1b3f4a137\" (UID: \"33e4ef6f-578b-4149-8d23-fcf1b3f4a137\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.185403 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c764da5f-6cfc-48aa-8a10-bf6bd97a226f-operator-scripts\") pod \"c764da5f-6cfc-48aa-8a10-bf6bd97a226f\" (UID: \"c764da5f-6cfc-48aa-8a10-bf6bd97a226f\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.185422 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xbc6\" (UniqueName: \"kubernetes.io/projected/c764da5f-6cfc-48aa-8a10-bf6bd97a226f-kube-api-access-5xbc6\") pod \"c764da5f-6cfc-48aa-8a10-bf6bd97a226f\" (UID: \"c764da5f-6cfc-48aa-8a10-bf6bd97a226f\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.185485 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33e4ef6f-578b-4149-8d23-fcf1b3f4a137-operator-scripts\") pod \"33e4ef6f-578b-4149-8d23-fcf1b3f4a137\" (UID: \"33e4ef6f-578b-4149-8d23-fcf1b3f4a137\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.185975 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9922aaf4-6776-4e6c-ac62-82c43bef4697-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.185992 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxmp9\" (UniqueName: \"kubernetes.io/projected/9922aaf4-6776-4e6c-ac62-82c43bef4697-kube-api-access-rxmp9\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.186360 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33e4ef6f-578b-4149-8d23-fcf1b3f4a137-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "33e4ef6f-578b-4149-8d23-fcf1b3f4a137" (UID: "33e4ef6f-578b-4149-8d23-fcf1b3f4a137"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.186899 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4c10-account-create-update-kmph9" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.187343 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c764da5f-6cfc-48aa-8a10-bf6bd97a226f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c764da5f-6cfc-48aa-8a10-bf6bd97a226f" (UID: "c764da5f-6cfc-48aa-8a10-bf6bd97a226f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.190551 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33e4ef6f-578b-4149-8d23-fcf1b3f4a137-kube-api-access-stbrp" (OuterVolumeSpecName: "kube-api-access-stbrp") pod "33e4ef6f-578b-4149-8d23-fcf1b3f4a137" (UID: "33e4ef6f-578b-4149-8d23-fcf1b3f4a137"). InnerVolumeSpecName "kube-api-access-stbrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.191876 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dvqcv" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.194947 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c764da5f-6cfc-48aa-8a10-bf6bd97a226f-kube-api-access-5xbc6" (OuterVolumeSpecName: "kube-api-access-5xbc6") pod "c764da5f-6cfc-48aa-8a10-bf6bd97a226f" (UID: "c764da5f-6cfc-48aa-8a10-bf6bd97a226f"). InnerVolumeSpecName "kube-api-access-5xbc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.201762 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8791-account-create-update-j27h4" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.287083 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0bb5aee-14d4-4023-bb79-fdd1824aae0e-operator-scripts\") pod \"a0bb5aee-14d4-4023-bb79-fdd1824aae0e\" (UID: \"a0bb5aee-14d4-4023-bb79-fdd1824aae0e\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.287173 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4nxt\" (UniqueName: \"kubernetes.io/projected/7f9006f4-9196-4a39-8c4d-10c4344a3b1f-kube-api-access-p4nxt\") pod \"7f9006f4-9196-4a39-8c4d-10c4344a3b1f\" (UID: \"7f9006f4-9196-4a39-8c4d-10c4344a3b1f\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.287216 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnmxl\" (UniqueName: \"kubernetes.io/projected/e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05-kube-api-access-tnmxl\") pod \"e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05\" (UID: \"e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.287243 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f9006f4-9196-4a39-8c4d-10c4344a3b1f-operator-scripts\") pod \"7f9006f4-9196-4a39-8c4d-10c4344a3b1f\" (UID: \"7f9006f4-9196-4a39-8c4d-10c4344a3b1f\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.287279 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05-operator-scripts\") pod \"e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05\" (UID: \"e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.287305 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c11795b-a2e2-4906-b3fd-3048ec29a764-operator-scripts\") pod \"1c11795b-a2e2-4906-b3fd-3048ec29a764\" (UID: \"1c11795b-a2e2-4906-b3fd-3048ec29a764\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.287326 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcqd4\" (UniqueName: \"kubernetes.io/projected/12142209-d074-426f-8e5d-d9e4acae21a0-kube-api-access-xcqd4\") pod \"12142209-d074-426f-8e5d-d9e4acae21a0\" (UID: \"12142209-d074-426f-8e5d-d9e4acae21a0\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.287397 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htr7l\" (UniqueName: \"kubernetes.io/projected/1c11795b-a2e2-4906-b3fd-3048ec29a764-kube-api-access-htr7l\") pod \"1c11795b-a2e2-4906-b3fd-3048ec29a764\" (UID: \"1c11795b-a2e2-4906-b3fd-3048ec29a764\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.287459 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12142209-d074-426f-8e5d-d9e4acae21a0-operator-scripts\") pod \"12142209-d074-426f-8e5d-d9e4acae21a0\" (UID: \"12142209-d074-426f-8e5d-d9e4acae21a0\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.287561 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbf8n\" (UniqueName: \"kubernetes.io/projected/a0bb5aee-14d4-4023-bb79-fdd1824aae0e-kube-api-access-cbf8n\") pod \"a0bb5aee-14d4-4023-bb79-fdd1824aae0e\" (UID: \"a0bb5aee-14d4-4023-bb79-fdd1824aae0e\") " Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.288174 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05" (UID: "e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.288387 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c11795b-a2e2-4906-b3fd-3048ec29a764-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1c11795b-a2e2-4906-b3fd-3048ec29a764" (UID: "1c11795b-a2e2-4906-b3fd-3048ec29a764"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.288774 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stbrp\" (UniqueName: \"kubernetes.io/projected/33e4ef6f-578b-4149-8d23-fcf1b3f4a137-kube-api-access-stbrp\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.288821 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c764da5f-6cfc-48aa-8a10-bf6bd97a226f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.288841 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xbc6\" (UniqueName: \"kubernetes.io/projected/c764da5f-6cfc-48aa-8a10-bf6bd97a226f-kube-api-access-5xbc6\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.288860 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33e4ef6f-578b-4149-8d23-fcf1b3f4a137-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.288706 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f9006f4-9196-4a39-8c4d-10c4344a3b1f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7f9006f4-9196-4a39-8c4d-10c4344a3b1f" (UID: "7f9006f4-9196-4a39-8c4d-10c4344a3b1f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.289101 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12142209-d074-426f-8e5d-d9e4acae21a0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "12142209-d074-426f-8e5d-d9e4acae21a0" (UID: "12142209-d074-426f-8e5d-d9e4acae21a0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.289400 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0bb5aee-14d4-4023-bb79-fdd1824aae0e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a0bb5aee-14d4-4023-bb79-fdd1824aae0e" (UID: "a0bb5aee-14d4-4023-bb79-fdd1824aae0e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.290464 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05-kube-api-access-tnmxl" (OuterVolumeSpecName: "kube-api-access-tnmxl") pod "e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05" (UID: "e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05"). InnerVolumeSpecName "kube-api-access-tnmxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.291051 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f9006f4-9196-4a39-8c4d-10c4344a3b1f-kube-api-access-p4nxt" (OuterVolumeSpecName: "kube-api-access-p4nxt") pod "7f9006f4-9196-4a39-8c4d-10c4344a3b1f" (UID: "7f9006f4-9196-4a39-8c4d-10c4344a3b1f"). InnerVolumeSpecName "kube-api-access-p4nxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.291318 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c11795b-a2e2-4906-b3fd-3048ec29a764-kube-api-access-htr7l" (OuterVolumeSpecName: "kube-api-access-htr7l") pod "1c11795b-a2e2-4906-b3fd-3048ec29a764" (UID: "1c11795b-a2e2-4906-b3fd-3048ec29a764"). InnerVolumeSpecName "kube-api-access-htr7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.291637 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12142209-d074-426f-8e5d-d9e4acae21a0-kube-api-access-xcqd4" (OuterVolumeSpecName: "kube-api-access-xcqd4") pod "12142209-d074-426f-8e5d-d9e4acae21a0" (UID: "12142209-d074-426f-8e5d-d9e4acae21a0"). InnerVolumeSpecName "kube-api-access-xcqd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.292713 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0bb5aee-14d4-4023-bb79-fdd1824aae0e-kube-api-access-cbf8n" (OuterVolumeSpecName: "kube-api-access-cbf8n") pod "a0bb5aee-14d4-4023-bb79-fdd1824aae0e" (UID: "a0bb5aee-14d4-4023-bb79-fdd1824aae0e"). InnerVolumeSpecName "kube-api-access-cbf8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.391364 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12142209-d074-426f-8e5d-d9e4acae21a0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.391403 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbf8n\" (UniqueName: \"kubernetes.io/projected/a0bb5aee-14d4-4023-bb79-fdd1824aae0e-kube-api-access-cbf8n\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.391414 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a0bb5aee-14d4-4023-bb79-fdd1824aae0e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.391423 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4nxt\" (UniqueName: \"kubernetes.io/projected/7f9006f4-9196-4a39-8c4d-10c4344a3b1f-kube-api-access-p4nxt\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.391431 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnmxl\" (UniqueName: \"kubernetes.io/projected/e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05-kube-api-access-tnmxl\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.391439 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f9006f4-9196-4a39-8c4d-10c4344a3b1f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.391447 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.391455 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c11795b-a2e2-4906-b3fd-3048ec29a764-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.391463 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcqd4\" (UniqueName: \"kubernetes.io/projected/12142209-d074-426f-8e5d-d9e4acae21a0-kube-api-access-xcqd4\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.391472 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htr7l\" (UniqueName: \"kubernetes.io/projected/1c11795b-a2e2-4906-b3fd-3048ec29a764-kube-api-access-htr7l\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.499853 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-4c10-account-create-update-kmph9" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.499842 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-4c10-account-create-update-kmph9" event={"ID":"1c11795b-a2e2-4906-b3fd-3048ec29a764","Type":"ContainerDied","Data":"783e13ee39326019cfff3e54328d32a27ac69ad161c4be419d7b0b418c8e341f"} Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.500571 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="783e13ee39326019cfff3e54328d32a27ac69ad161c4be419d7b0b418c8e341f" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.503961 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9g2b5" event={"ID":"c764da5f-6cfc-48aa-8a10-bf6bd97a226f","Type":"ContainerDied","Data":"d422ba352d956106286ca781e5a02094801b4bdbd7315ec4f77ccf8b2f0dc340"} Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.504000 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9g2b5" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.504011 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d422ba352d956106286ca781e5a02094801b4bdbd7315ec4f77ccf8b2f0dc340" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.506322 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-355c-account-create-update-d5b2t" event={"ID":"33e4ef6f-578b-4149-8d23-fcf1b3f4a137","Type":"ContainerDied","Data":"f77a0300f4f70f744d9a8a4de7aedab343e6199aa089b6d950c0932950d214d0"} Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.506357 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f77a0300f4f70f744d9a8a4de7aedab343e6199aa089b6d950c0932950d214d0" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.506406 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-355c-account-create-update-d5b2t" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.508607 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-dvqcv" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.508625 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-dvqcv" event={"ID":"7f9006f4-9196-4a39-8c4d-10c4344a3b1f","Type":"ContainerDied","Data":"2f291e130e58a5beabfcb89379832ccb0d526cf2c51c7236e190d86be6e226d6"} Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.508669 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f291e130e58a5beabfcb89379832ccb0d526cf2c51c7236e190d86be6e226d6" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.511696 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8791-account-create-update-j27h4" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.511692 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8791-account-create-update-j27h4" event={"ID":"12142209-d074-426f-8e5d-d9e4acae21a0","Type":"ContainerDied","Data":"0615344d80c71e58ce944d2b37ef5cf4722267f310ee747b8795ee904a285e02"} Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.511993 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0615344d80c71e58ce944d2b37ef5cf4722267f310ee747b8795ee904a285e02" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.515231 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-4d9f-account-create-update-bc8h8" event={"ID":"e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05","Type":"ContainerDied","Data":"d3b5475f60b0069b3845972199410debe6b71383b9efd2a9e435647268ebfdb4"} Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.515260 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3b5475f60b0069b3845972199410debe6b71383b9efd2a9e435647268ebfdb4" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.515281 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-4d9f-account-create-update-bc8h8" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.519066 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-create-484hp" event={"ID":"a0bb5aee-14d4-4023-bb79-fdd1824aae0e","Type":"ContainerDied","Data":"01e33d7675021eba461367e1f50b83f3c60ba7664bcb505845c569a1c9dc6df3"} Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.519237 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01e33d7675021eba461367e1f50b83f3c60ba7664bcb505845c569a1c9dc6df3" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.519458 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-create-484hp" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.526008 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8g74n" event={"ID":"9922aaf4-6776-4e6c-ac62-82c43bef4697","Type":"ContainerDied","Data":"1aac1696eb44fdb0163d26ae614d375a1fc4e853072fe435ca0fcff3fdce8a75"} Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.526075 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1aac1696eb44fdb0163d26ae614d375a1fc4e853072fe435ca0fcff3fdce8a75" Dec 02 01:55:27 crc kubenswrapper[4884]: I1202 01:55:27.526151 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8g74n" Dec 02 01:55:29 crc kubenswrapper[4884]: I1202 01:55:29.019308 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 02 01:55:29 crc kubenswrapper[4884]: I1202 01:55:29.020613 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:29 crc kubenswrapper[4884]: E1202 01:55:29.020780 4884 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 02 01:55:29 crc kubenswrapper[4884]: E1202 01:55:29.020799 4884 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 02 01:55:29 crc kubenswrapper[4884]: E1202 01:55:29.020847 4884 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift podName:8dc2a1f5-a7a6-4ce5-a711-77455efe2224 nodeName:}" failed. No retries permitted until 2025-12-02 01:55:45.020831481 +0000 UTC m=+1161.696668365 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift") pod "swift-storage-0" (UID: "8dc2a1f5-a7a6-4ce5-a711-77455efe2224") : configmap "swift-ring-files" not found Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.319579 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-mrh2r"] Dec 02 01:55:30 crc kubenswrapper[4884]: E1202 01:55:30.320205 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d" containerName="init" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320218 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d" containerName="init" Dec 02 01:55:30 crc kubenswrapper[4884]: E1202 01:55:30.320228 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f9006f4-9196-4a39-8c4d-10c4344a3b1f" containerName="mariadb-database-create" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320234 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f9006f4-9196-4a39-8c4d-10c4344a3b1f" containerName="mariadb-database-create" Dec 02 01:55:30 crc kubenswrapper[4884]: E1202 01:55:30.320251 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c764da5f-6cfc-48aa-8a10-bf6bd97a226f" containerName="mariadb-database-create" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320257 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="c764da5f-6cfc-48aa-8a10-bf6bd97a226f" containerName="mariadb-database-create" Dec 02 01:55:30 crc kubenswrapper[4884]: E1202 01:55:30.320271 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12142209-d074-426f-8e5d-d9e4acae21a0" containerName="mariadb-account-create-update" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320276 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="12142209-d074-426f-8e5d-d9e4acae21a0" containerName="mariadb-account-create-update" Dec 02 01:55:30 crc kubenswrapper[4884]: E1202 01:55:30.320285 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9922aaf4-6776-4e6c-ac62-82c43bef4697" containerName="mariadb-database-create" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320291 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="9922aaf4-6776-4e6c-ac62-82c43bef4697" containerName="mariadb-database-create" Dec 02 01:55:30 crc kubenswrapper[4884]: E1202 01:55:30.320304 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0bb5aee-14d4-4023-bb79-fdd1824aae0e" containerName="mariadb-database-create" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320310 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0bb5aee-14d4-4023-bb79-fdd1824aae0e" containerName="mariadb-database-create" Dec 02 01:55:30 crc kubenswrapper[4884]: E1202 01:55:30.320320 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05" containerName="mariadb-account-create-update" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320326 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05" containerName="mariadb-account-create-update" Dec 02 01:55:30 crc kubenswrapper[4884]: E1202 01:55:30.320340 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d" containerName="dnsmasq-dns" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320346 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d" containerName="dnsmasq-dns" Dec 02 01:55:30 crc kubenswrapper[4884]: E1202 01:55:30.320355 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c11795b-a2e2-4906-b3fd-3048ec29a764" containerName="mariadb-account-create-update" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320360 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c11795b-a2e2-4906-b3fd-3048ec29a764" containerName="mariadb-account-create-update" Dec 02 01:55:30 crc kubenswrapper[4884]: E1202 01:55:30.320371 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33e4ef6f-578b-4149-8d23-fcf1b3f4a137" containerName="mariadb-account-create-update" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320377 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="33e4ef6f-578b-4149-8d23-fcf1b3f4a137" containerName="mariadb-account-create-update" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320525 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c11795b-a2e2-4906-b3fd-3048ec29a764" containerName="mariadb-account-create-update" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320548 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05" containerName="mariadb-account-create-update" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320555 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f9006f4-9196-4a39-8c4d-10c4344a3b1f" containerName="mariadb-database-create" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320567 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="33e4ef6f-578b-4149-8d23-fcf1b3f4a137" containerName="mariadb-account-create-update" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320580 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0bb5aee-14d4-4023-bb79-fdd1824aae0e" containerName="mariadb-database-create" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320590 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="12142209-d074-426f-8e5d-d9e4acae21a0" containerName="mariadb-account-create-update" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320607 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="c764da5f-6cfc-48aa-8a10-bf6bd97a226f" containerName="mariadb-database-create" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320619 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="01126ef4-d34f-4aa4-b8c4-6c96a27f6c1d" containerName="dnsmasq-dns" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.320631 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="9922aaf4-6776-4e6c-ac62-82c43bef4697" containerName="mariadb-database-create" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.321169 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.324685 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.326541 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-rvbnc" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.335123 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-mrh2r"] Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.447391 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wrfk\" (UniqueName: \"kubernetes.io/projected/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-kube-api-access-9wrfk\") pod \"glance-db-sync-mrh2r\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.447489 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-combined-ca-bundle\") pod \"glance-db-sync-mrh2r\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.447531 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-config-data\") pod \"glance-db-sync-mrh2r\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.447556 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-db-sync-config-data\") pod \"glance-db-sync-mrh2r\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.551343 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wrfk\" (UniqueName: \"kubernetes.io/projected/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-kube-api-access-9wrfk\") pod \"glance-db-sync-mrh2r\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.551566 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-combined-ca-bundle\") pod \"glance-db-sync-mrh2r\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.551679 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-config-data\") pod \"glance-db-sync-mrh2r\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.551732 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-db-sync-config-data\") pod \"glance-db-sync-mrh2r\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.556093 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c8bb2562-5d06-44d1-bbd0-8848d92f18d2","Type":"ContainerStarted","Data":"8a72e4291e35b4fb3bca6fa907132a6a3137ecf42f9d5a1e0c06bd5bf7d4c831"} Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.557911 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-db-sync-config-data\") pod \"glance-db-sync-mrh2r\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.559083 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-combined-ca-bundle\") pod \"glance-db-sync-mrh2r\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.560126 4884 generic.go:334] "Generic (PLEG): container finished" podID="151b4ede-9694-42d6-8051-6309de26a4e4" containerID="fdc7b71ec90bc36daeab689156c99d0836cfd20761769249ac79beee976a010f" exitCode=0 Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.560171 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-l2j2h" event={"ID":"151b4ede-9694-42d6-8051-6309de26a4e4","Type":"ContainerDied","Data":"fdc7b71ec90bc36daeab689156c99d0836cfd20761769249ac79beee976a010f"} Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.563722 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-config-data\") pod \"glance-db-sync-mrh2r\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.578455 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wrfk\" (UniqueName: \"kubernetes.io/projected/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-kube-api-access-9wrfk\") pod \"glance-db-sync-mrh2r\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.590051 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=11.215799434000001 podStartE2EDuration="49.590026214s" podCreationTimestamp="2025-12-02 01:54:41 +0000 UTC" firstStartedPulling="2025-12-02 01:54:51.717288411 +0000 UTC m=+1108.393125295" lastFinishedPulling="2025-12-02 01:55:30.091515181 +0000 UTC m=+1146.767352075" observedRunningTime="2025-12-02 01:55:30.578519465 +0000 UTC m=+1147.254356359" watchObservedRunningTime="2025-12-02 01:55:30.590026214 +0000 UTC m=+1147.265863128" Dec 02 01:55:30 crc kubenswrapper[4884]: I1202 01:55:30.634660 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:31 crc kubenswrapper[4884]: I1202 01:55:31.232048 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-mrh2r"] Dec 02 01:55:31 crc kubenswrapper[4884]: W1202 01:55:31.236386 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83c3e77b_e1e2_4ee7_bd7c_601e4f90ff5e.slice/crio-2d3e14c4cf3fd1403b1db225c5fc85b2a852681b84a51dd4233f9661ca8c3cb4 WatchSource:0}: Error finding container 2d3e14c4cf3fd1403b1db225c5fc85b2a852681b84a51dd4233f9661ca8c3cb4: Status 404 returned error can't find the container with id 2d3e14c4cf3fd1403b1db225c5fc85b2a852681b84a51dd4233f9661ca8c3cb4 Dec 02 01:55:31 crc kubenswrapper[4884]: I1202 01:55:31.578633 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mrh2r" event={"ID":"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e","Type":"ContainerStarted","Data":"2d3e14c4cf3fd1403b1db225c5fc85b2a852681b84a51dd4233f9661ca8c3cb4"} Dec 02 01:55:31 crc kubenswrapper[4884]: I1202 01:55:31.928810 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:31 crc kubenswrapper[4884]: I1202 01:55:31.982872 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/151b4ede-9694-42d6-8051-6309de26a4e4-ring-data-devices\") pod \"151b4ede-9694-42d6-8051-6309de26a4e4\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " Dec 02 01:55:31 crc kubenswrapper[4884]: I1202 01:55:31.982937 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/151b4ede-9694-42d6-8051-6309de26a4e4-scripts\") pod \"151b4ede-9694-42d6-8051-6309de26a4e4\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " Dec 02 01:55:31 crc kubenswrapper[4884]: I1202 01:55:31.983006 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-swiftconf\") pod \"151b4ede-9694-42d6-8051-6309de26a4e4\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " Dec 02 01:55:31 crc kubenswrapper[4884]: I1202 01:55:31.983033 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-dispersionconf\") pod \"151b4ede-9694-42d6-8051-6309de26a4e4\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " Dec 02 01:55:31 crc kubenswrapper[4884]: I1202 01:55:31.983094 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-combined-ca-bundle\") pod \"151b4ede-9694-42d6-8051-6309de26a4e4\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " Dec 02 01:55:31 crc kubenswrapper[4884]: I1202 01:55:31.983181 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/151b4ede-9694-42d6-8051-6309de26a4e4-etc-swift\") pod \"151b4ede-9694-42d6-8051-6309de26a4e4\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " Dec 02 01:55:31 crc kubenswrapper[4884]: I1202 01:55:31.983200 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lld2z\" (UniqueName: \"kubernetes.io/projected/151b4ede-9694-42d6-8051-6309de26a4e4-kube-api-access-lld2z\") pod \"151b4ede-9694-42d6-8051-6309de26a4e4\" (UID: \"151b4ede-9694-42d6-8051-6309de26a4e4\") " Dec 02 01:55:31 crc kubenswrapper[4884]: I1202 01:55:31.984681 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/151b4ede-9694-42d6-8051-6309de26a4e4-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "151b4ede-9694-42d6-8051-6309de26a4e4" (UID: "151b4ede-9694-42d6-8051-6309de26a4e4"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:31 crc kubenswrapper[4884]: I1202 01:55:31.985493 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/151b4ede-9694-42d6-8051-6309de26a4e4-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "151b4ede-9694-42d6-8051-6309de26a4e4" (UID: "151b4ede-9694-42d6-8051-6309de26a4e4"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:55:31 crc kubenswrapper[4884]: I1202 01:55:31.990970 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/151b4ede-9694-42d6-8051-6309de26a4e4-kube-api-access-lld2z" (OuterVolumeSpecName: "kube-api-access-lld2z") pod "151b4ede-9694-42d6-8051-6309de26a4e4" (UID: "151b4ede-9694-42d6-8051-6309de26a4e4"). InnerVolumeSpecName "kube-api-access-lld2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:31 crc kubenswrapper[4884]: I1202 01:55:31.993332 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "151b4ede-9694-42d6-8051-6309de26a4e4" (UID: "151b4ede-9694-42d6-8051-6309de26a4e4"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.015403 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "151b4ede-9694-42d6-8051-6309de26a4e4" (UID: "151b4ede-9694-42d6-8051-6309de26a4e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.019788 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/151b4ede-9694-42d6-8051-6309de26a4e4-scripts" (OuterVolumeSpecName: "scripts") pod "151b4ede-9694-42d6-8051-6309de26a4e4" (UID: "151b4ede-9694-42d6-8051-6309de26a4e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.027068 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "151b4ede-9694-42d6-8051-6309de26a4e4" (UID: "151b4ede-9694-42d6-8051-6309de26a4e4"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.084618 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.084660 4884 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/151b4ede-9694-42d6-8051-6309de26a4e4-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.084672 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lld2z\" (UniqueName: \"kubernetes.io/projected/151b4ede-9694-42d6-8051-6309de26a4e4-kube-api-access-lld2z\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.084685 4884 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/151b4ede-9694-42d6-8051-6309de26a4e4-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.084698 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/151b4ede-9694-42d6-8051-6309de26a4e4-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.084708 4884 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.084720 4884 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/151b4ede-9694-42d6-8051-6309de26a4e4-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.574883 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-9wkcq" podUID="cdabd8d6-7f1f-4629-b6cb-1be182bd9b66" containerName="ovn-controller" probeResult="failure" output=< Dec 02 01:55:32 crc kubenswrapper[4884]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 02 01:55:32 crc kubenswrapper[4884]: > Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.590692 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-l2j2h" event={"ID":"151b4ede-9694-42d6-8051-6309de26a4e4","Type":"ContainerDied","Data":"2dbe7c38a965077060c06ab05c2b00a16fafab5808d9b92887b2cd9f722c8b7b"} Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.590733 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dbe7c38a965077060c06ab05c2b00a16fafab5808d9b92887b2cd9f722c8b7b" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.590827 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-l2j2h" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.599175 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.617524 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-bdv77" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.834509 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-9wkcq-config-rqjpj"] Dec 02 01:55:32 crc kubenswrapper[4884]: E1202 01:55:32.834961 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="151b4ede-9694-42d6-8051-6309de26a4e4" containerName="swift-ring-rebalance" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.834982 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="151b4ede-9694-42d6-8051-6309de26a4e4" containerName="swift-ring-rebalance" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.835209 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="151b4ede-9694-42d6-8051-6309de26a4e4" containerName="swift-ring-rebalance" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.835990 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.838126 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.865354 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9wkcq-config-rqjpj"] Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.909050 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-log-ovn\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.909122 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcpt4\" (UniqueName: \"kubernetes.io/projected/efe89ca4-1e2c-4531-b52c-f99028ef6d14-kube-api-access-hcpt4\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.909186 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/efe89ca4-1e2c-4531-b52c-f99028ef6d14-additional-scripts\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.909226 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efe89ca4-1e2c-4531-b52c-f99028ef6d14-scripts\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.909251 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-run-ovn\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:32 crc kubenswrapper[4884]: I1202 01:55:32.909510 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-run\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.010867 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/efe89ca4-1e2c-4531-b52c-f99028ef6d14-additional-scripts\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.010932 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efe89ca4-1e2c-4531-b52c-f99028ef6d14-scripts\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.010954 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-run-ovn\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.011003 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-run\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.011062 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-log-ovn\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.011106 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcpt4\" (UniqueName: \"kubernetes.io/projected/efe89ca4-1e2c-4531-b52c-f99028ef6d14-kube-api-access-hcpt4\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.011786 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-log-ovn\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.011725 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-run\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.012025 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/efe89ca4-1e2c-4531-b52c-f99028ef6d14-additional-scripts\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.014009 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efe89ca4-1e2c-4531-b52c-f99028ef6d14-scripts\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.014068 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-run-ovn\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.027129 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcpt4\" (UniqueName: \"kubernetes.io/projected/efe89ca4-1e2c-4531-b52c-f99028ef6d14-kube-api-access-hcpt4\") pod \"ovn-controller-9wkcq-config-rqjpj\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.155035 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.187695 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.601224 4884 generic.go:334] "Generic (PLEG): container finished" podID="f32d61be-992e-4a09-9663-33d3d59779e9" containerID="39010f45dce6a3ffe62ec61181fd5ae24a376cc18ac028cb8715cf1211681fc2" exitCode=0 Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.601283 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f32d61be-992e-4a09-9663-33d3d59779e9","Type":"ContainerDied","Data":"39010f45dce6a3ffe62ec61181fd5ae24a376cc18ac028cb8715cf1211681fc2"} Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.602917 4884 generic.go:334] "Generic (PLEG): container finished" podID="27182112-178e-415c-9af9-f7fbf58d8120" containerID="b9dfb8cc8de08b7d246a2396788751343d208c51ed83bbee65b5010ac935f251" exitCode=0 Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.603683 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"27182112-178e-415c-9af9-f7fbf58d8120","Type":"ContainerDied","Data":"b9dfb8cc8de08b7d246a2396788751343d208c51ed83bbee65b5010ac935f251"} Dec 02 01:55:33 crc kubenswrapper[4884]: I1202 01:55:33.691514 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9wkcq-config-rqjpj"] Dec 02 01:55:34 crc kubenswrapper[4884]: I1202 01:55:34.617360 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f32d61be-992e-4a09-9663-33d3d59779e9","Type":"ContainerStarted","Data":"667241041ac23c701da96842218f42323b58836f1801238ed1c63c9a447b69d9"} Dec 02 01:55:34 crc kubenswrapper[4884]: I1202 01:55:34.617830 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 02 01:55:34 crc kubenswrapper[4884]: I1202 01:55:34.619732 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"27182112-178e-415c-9af9-f7fbf58d8120","Type":"ContainerStarted","Data":"37cfb917f66312e86b6bacb7e41b56101104db52b4ae5ab3ec242ca5f852baf0"} Dec 02 01:55:34 crc kubenswrapper[4884]: I1202 01:55:34.619935 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:55:34 crc kubenswrapper[4884]: I1202 01:55:34.626949 4884 generic.go:334] "Generic (PLEG): container finished" podID="efe89ca4-1e2c-4531-b52c-f99028ef6d14" containerID="2c260c8245a4227ce0886eb4629815934d6639a7790618f122cc64e375a82684" exitCode=0 Dec 02 01:55:34 crc kubenswrapper[4884]: I1202 01:55:34.627019 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9wkcq-config-rqjpj" event={"ID":"efe89ca4-1e2c-4531-b52c-f99028ef6d14","Type":"ContainerDied","Data":"2c260c8245a4227ce0886eb4629815934d6639a7790618f122cc64e375a82684"} Dec 02 01:55:34 crc kubenswrapper[4884]: I1202 01:55:34.627054 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9wkcq-config-rqjpj" event={"ID":"efe89ca4-1e2c-4531-b52c-f99028ef6d14","Type":"ContainerStarted","Data":"31e23aa38e833d6d73a52be99bbef0074d63a0ab4141ad1e47941fca8efd14f8"} Dec 02 01:55:34 crc kubenswrapper[4884]: I1202 01:55:34.645508 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=51.935544279 podStartE2EDuration="1m0.645492546s" podCreationTimestamp="2025-12-02 01:54:34 +0000 UTC" firstStartedPulling="2025-12-02 01:54:51.166480332 +0000 UTC m=+1107.842317206" lastFinishedPulling="2025-12-02 01:54:59.876428589 +0000 UTC m=+1116.552265473" observedRunningTime="2025-12-02 01:55:34.645089706 +0000 UTC m=+1151.320926620" watchObservedRunningTime="2025-12-02 01:55:34.645492546 +0000 UTC m=+1151.321329430" Dec 02 01:55:34 crc kubenswrapper[4884]: I1202 01:55:34.678175 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=51.555307493 podStartE2EDuration="59.678157598s" podCreationTimestamp="2025-12-02 01:54:35 +0000 UTC" firstStartedPulling="2025-12-02 01:54:50.921464777 +0000 UTC m=+1107.597301661" lastFinishedPulling="2025-12-02 01:54:59.044314842 +0000 UTC m=+1115.720151766" observedRunningTime="2025-12-02 01:55:34.673637785 +0000 UTC m=+1151.349474669" watchObservedRunningTime="2025-12-02 01:55:34.678157598 +0000 UTC m=+1151.353994482" Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.024387 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.063277 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcpt4\" (UniqueName: \"kubernetes.io/projected/efe89ca4-1e2c-4531-b52c-f99028ef6d14-kube-api-access-hcpt4\") pod \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.063331 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-log-ovn\") pod \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.063354 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/efe89ca4-1e2c-4531-b52c-f99028ef6d14-additional-scripts\") pod \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.063380 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-run-ovn\") pod \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.063400 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efe89ca4-1e2c-4531-b52c-f99028ef6d14-scripts\") pod \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.063450 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-run\") pod \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\" (UID: \"efe89ca4-1e2c-4531-b52c-f99028ef6d14\") " Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.063815 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-run" (OuterVolumeSpecName: "var-run") pod "efe89ca4-1e2c-4531-b52c-f99028ef6d14" (UID: "efe89ca4-1e2c-4531-b52c-f99028ef6d14"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.063848 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "efe89ca4-1e2c-4531-b52c-f99028ef6d14" (UID: "efe89ca4-1e2c-4531-b52c-f99028ef6d14"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.064391 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "efe89ca4-1e2c-4531-b52c-f99028ef6d14" (UID: "efe89ca4-1e2c-4531-b52c-f99028ef6d14"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.064490 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efe89ca4-1e2c-4531-b52c-f99028ef6d14-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "efe89ca4-1e2c-4531-b52c-f99028ef6d14" (UID: "efe89ca4-1e2c-4531-b52c-f99028ef6d14"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.064908 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efe89ca4-1e2c-4531-b52c-f99028ef6d14-scripts" (OuterVolumeSpecName: "scripts") pod "efe89ca4-1e2c-4531-b52c-f99028ef6d14" (UID: "efe89ca4-1e2c-4531-b52c-f99028ef6d14"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.069944 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efe89ca4-1e2c-4531-b52c-f99028ef6d14-kube-api-access-hcpt4" (OuterVolumeSpecName: "kube-api-access-hcpt4") pod "efe89ca4-1e2c-4531-b52c-f99028ef6d14" (UID: "efe89ca4-1e2c-4531-b52c-f99028ef6d14"). InnerVolumeSpecName "kube-api-access-hcpt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.164782 4884 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.165051 4884 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/efe89ca4-1e2c-4531-b52c-f99028ef6d14-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.165061 4884 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.165070 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/efe89ca4-1e2c-4531-b52c-f99028ef6d14-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.165080 4884 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/efe89ca4-1e2c-4531-b52c-f99028ef6d14-var-run\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.165088 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcpt4\" (UniqueName: \"kubernetes.io/projected/efe89ca4-1e2c-4531-b52c-f99028ef6d14-kube-api-access-hcpt4\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.645880 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9wkcq-config-rqjpj" event={"ID":"efe89ca4-1e2c-4531-b52c-f99028ef6d14","Type":"ContainerDied","Data":"31e23aa38e833d6d73a52be99bbef0074d63a0ab4141ad1e47941fca8efd14f8"} Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.645916 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31e23aa38e833d6d73a52be99bbef0074d63a0ab4141ad1e47941fca8efd14f8" Dec 02 01:55:36 crc kubenswrapper[4884]: I1202 01:55:36.645973 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9wkcq-config-rqjpj" Dec 02 01:55:37 crc kubenswrapper[4884]: I1202 01:55:37.122112 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-9wkcq-config-rqjpj"] Dec 02 01:55:37 crc kubenswrapper[4884]: I1202 01:55:37.128934 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-9wkcq-config-rqjpj"] Dec 02 01:55:37 crc kubenswrapper[4884]: I1202 01:55:37.582092 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-9wkcq" Dec 02 01:55:37 crc kubenswrapper[4884]: I1202 01:55:37.632936 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efe89ca4-1e2c-4531-b52c-f99028ef6d14" path="/var/lib/kubelet/pods/efe89ca4-1e2c-4531-b52c-f99028ef6d14/volumes" Dec 02 01:55:43 crc kubenswrapper[4884]: I1202 01:55:43.212852 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:43 crc kubenswrapper[4884]: I1202 01:55:43.223178 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:43 crc kubenswrapper[4884]: I1202 01:55:43.705840 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:45 crc kubenswrapper[4884]: I1202 01:55:45.045701 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:45 crc kubenswrapper[4884]: I1202 01:55:45.052525 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/8dc2a1f5-a7a6-4ce5-a711-77455efe2224-etc-swift\") pod \"swift-storage-0\" (UID: \"8dc2a1f5-a7a6-4ce5-a711-77455efe2224\") " pod="openstack/swift-storage-0" Dec 02 01:55:45 crc kubenswrapper[4884]: I1202 01:55:45.123849 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.173134 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.189586 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.190144 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerName="prometheus" containerID="cri-o://05e715ded10f8d73e952217f675ad561ba62a82b52a6306b6dd3b7b674d8329e" gracePeriod=600 Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.190155 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerName="thanos-sidecar" containerID="cri-o://8a72e4291e35b4fb3bca6fa907132a6a3137ecf42f9d5a1e0c06bd5bf7d4c831" gracePeriod=600 Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.190172 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerName="config-reloader" containerID="cri-o://38a9fea55138df4195eab6c884265c3d5ff494f21607ba70544cff03fee07d66" gracePeriod=600 Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.445916 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.463271 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.629615 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-db-sync-pm699"] Dec 02 01:55:46 crc kubenswrapper[4884]: E1202 01:55:46.630598 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efe89ca4-1e2c-4531-b52c-f99028ef6d14" containerName="ovn-config" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.630617 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="efe89ca4-1e2c-4531-b52c-f99028ef6d14" containerName="ovn-config" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.630989 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="efe89ca4-1e2c-4531-b52c-f99028ef6d14" containerName="ovn-config" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.631935 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-pm699" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.635553 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-r5l44" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.636536 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-config-data" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.666911 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-pm699"] Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.707309 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-qfblm"] Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.708850 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qfblm" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.719899 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qfblm"] Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.732704 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"f75c1702bd2434862679cfa380f09bb741360b2a2a05ade2a25bbebce72b58fc"} Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.746218 4884 generic.go:334] "Generic (PLEG): container finished" podID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerID="8a72e4291e35b4fb3bca6fa907132a6a3137ecf42f9d5a1e0c06bd5bf7d4c831" exitCode=0 Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.746248 4884 generic.go:334] "Generic (PLEG): container finished" podID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerID="38a9fea55138df4195eab6c884265c3d5ff494f21607ba70544cff03fee07d66" exitCode=0 Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.746256 4884 generic.go:334] "Generic (PLEG): container finished" podID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerID="05e715ded10f8d73e952217f675ad561ba62a82b52a6306b6dd3b7b674d8329e" exitCode=0 Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.746277 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c8bb2562-5d06-44d1-bbd0-8848d92f18d2","Type":"ContainerDied","Data":"8a72e4291e35b4fb3bca6fa907132a6a3137ecf42f9d5a1e0c06bd5bf7d4c831"} Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.746302 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c8bb2562-5d06-44d1-bbd0-8848d92f18d2","Type":"ContainerDied","Data":"38a9fea55138df4195eab6c884265c3d5ff494f21607ba70544cff03fee07d66"} Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.746312 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c8bb2562-5d06-44d1-bbd0-8848d92f18d2","Type":"ContainerDied","Data":"05e715ded10f8d73e952217f675ad561ba62a82b52a6306b6dd3b7b674d8329e"} Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.781067 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-hnwz5"] Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.782110 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-hnwz5" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.801059 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-b26c-account-create-update-92x4v"] Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.802270 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b26c-account-create-update-92x4v" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.809831 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.823219 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-hnwz5"] Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.826433 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1749692-f3f0-4849-bdf5-73ce5bfec72a-operator-scripts\") pod \"cinder-db-create-qfblm\" (UID: \"e1749692-f3f0-4849-bdf5-73ce5bfec72a\") " pod="openstack/cinder-db-create-qfblm" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.826542 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-db-sync-config-data\") pod \"watcher-db-sync-pm699\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " pod="openstack/watcher-db-sync-pm699" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.826567 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh94v\" (UniqueName: \"kubernetes.io/projected/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-kube-api-access-lh94v\") pod \"watcher-db-sync-pm699\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " pod="openstack/watcher-db-sync-pm699" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.826597 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-config-data\") pod \"watcher-db-sync-pm699\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " pod="openstack/watcher-db-sync-pm699" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.826621 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-combined-ca-bundle\") pod \"watcher-db-sync-pm699\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " pod="openstack/watcher-db-sync-pm699" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.826653 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p4fb\" (UniqueName: \"kubernetes.io/projected/e1749692-f3f0-4849-bdf5-73ce5bfec72a-kube-api-access-5p4fb\") pod \"cinder-db-create-qfblm\" (UID: \"e1749692-f3f0-4849-bdf5-73ce5bfec72a\") " pod="openstack/cinder-db-create-qfblm" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.842478 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b26c-account-create-update-92x4v"] Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.889826 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-3040-account-create-update-mx8t8"] Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.890928 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3040-account-create-update-mx8t8" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.893972 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.907061 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3040-account-create-update-mx8t8"] Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.927513 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg5r9\" (UniqueName: \"kubernetes.io/projected/3d14133a-bc13-40b8-9c84-cad4a04bb32e-kube-api-access-tg5r9\") pod \"barbican-db-create-hnwz5\" (UID: \"3d14133a-bc13-40b8-9c84-cad4a04bb32e\") " pod="openstack/barbican-db-create-hnwz5" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.927562 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1749692-f3f0-4849-bdf5-73ce5bfec72a-operator-scripts\") pod \"cinder-db-create-qfblm\" (UID: \"e1749692-f3f0-4849-bdf5-73ce5bfec72a\") " pod="openstack/cinder-db-create-qfblm" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.927635 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-db-sync-config-data\") pod \"watcher-db-sync-pm699\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " pod="openstack/watcher-db-sync-pm699" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.927653 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh94v\" (UniqueName: \"kubernetes.io/projected/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-kube-api-access-lh94v\") pod \"watcher-db-sync-pm699\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " pod="openstack/watcher-db-sync-pm699" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.927683 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clrq5\" (UniqueName: \"kubernetes.io/projected/fb68c844-6d7b-4f52-9a57-1ba8a4603449-kube-api-access-clrq5\") pod \"barbican-b26c-account-create-update-92x4v\" (UID: \"fb68c844-6d7b-4f52-9a57-1ba8a4603449\") " pod="openstack/barbican-b26c-account-create-update-92x4v" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.927705 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-config-data\") pod \"watcher-db-sync-pm699\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " pod="openstack/watcher-db-sync-pm699" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.927721 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb68c844-6d7b-4f52-9a57-1ba8a4603449-operator-scripts\") pod \"barbican-b26c-account-create-update-92x4v\" (UID: \"fb68c844-6d7b-4f52-9a57-1ba8a4603449\") " pod="openstack/barbican-b26c-account-create-update-92x4v" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.927761 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-combined-ca-bundle\") pod \"watcher-db-sync-pm699\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " pod="openstack/watcher-db-sync-pm699" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.927790 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d14133a-bc13-40b8-9c84-cad4a04bb32e-operator-scripts\") pod \"barbican-db-create-hnwz5\" (UID: \"3d14133a-bc13-40b8-9c84-cad4a04bb32e\") " pod="openstack/barbican-db-create-hnwz5" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.927830 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p4fb\" (UniqueName: \"kubernetes.io/projected/e1749692-f3f0-4849-bdf5-73ce5bfec72a-kube-api-access-5p4fb\") pod \"cinder-db-create-qfblm\" (UID: \"e1749692-f3f0-4849-bdf5-73ce5bfec72a\") " pod="openstack/cinder-db-create-qfblm" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.928710 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1749692-f3f0-4849-bdf5-73ce5bfec72a-operator-scripts\") pod \"cinder-db-create-qfblm\" (UID: \"e1749692-f3f0-4849-bdf5-73ce5bfec72a\") " pod="openstack/cinder-db-create-qfblm" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.935348 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-db-sync-config-data\") pod \"watcher-db-sync-pm699\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " pod="openstack/watcher-db-sync-pm699" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.938252 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-combined-ca-bundle\") pod \"watcher-db-sync-pm699\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " pod="openstack/watcher-db-sync-pm699" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.938426 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-config-data\") pod \"watcher-db-sync-pm699\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " pod="openstack/watcher-db-sync-pm699" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.957500 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-6z2v4"] Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.963697 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6z2v4" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.964696 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p4fb\" (UniqueName: \"kubernetes.io/projected/e1749692-f3f0-4849-bdf5-73ce5bfec72a-kube-api-access-5p4fb\") pod \"cinder-db-create-qfblm\" (UID: \"e1749692-f3f0-4849-bdf5-73ce5bfec72a\") " pod="openstack/cinder-db-create-qfblm" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.973126 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.973360 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.973436 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-42pg8" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.973454 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.984813 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh94v\" (UniqueName: \"kubernetes.io/projected/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-kube-api-access-lh94v\") pod \"watcher-db-sync-pm699\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " pod="openstack/watcher-db-sync-pm699" Dec 02 01:55:46 crc kubenswrapper[4884]: I1202 01:55:46.987002 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-6z2v4"] Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.003360 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-zxwkn"] Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.004441 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zxwkn" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.028864 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb68c844-6d7b-4f52-9a57-1ba8a4603449-operator-scripts\") pod \"barbican-b26c-account-create-update-92x4v\" (UID: \"fb68c844-6d7b-4f52-9a57-1ba8a4603449\") " pod="openstack/barbican-b26c-account-create-update-92x4v" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.029141 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5722ebb-3612-4608-9768-6d8a95d86efe-operator-scripts\") pod \"cinder-3040-account-create-update-mx8t8\" (UID: \"f5722ebb-3612-4608-9768-6d8a95d86efe\") " pod="openstack/cinder-3040-account-create-update-mx8t8" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.029178 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d14133a-bc13-40b8-9c84-cad4a04bb32e-operator-scripts\") pod \"barbican-db-create-hnwz5\" (UID: \"3d14133a-bc13-40b8-9c84-cad4a04bb32e\") " pod="openstack/barbican-db-create-hnwz5" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.029210 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg5r9\" (UniqueName: \"kubernetes.io/projected/3d14133a-bc13-40b8-9c84-cad4a04bb32e-kube-api-access-tg5r9\") pod \"barbican-db-create-hnwz5\" (UID: \"3d14133a-bc13-40b8-9c84-cad4a04bb32e\") " pod="openstack/barbican-db-create-hnwz5" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.029239 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zstf6\" (UniqueName: \"kubernetes.io/projected/f5722ebb-3612-4608-9768-6d8a95d86efe-kube-api-access-zstf6\") pod \"cinder-3040-account-create-update-mx8t8\" (UID: \"f5722ebb-3612-4608-9768-6d8a95d86efe\") " pod="openstack/cinder-3040-account-create-update-mx8t8" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.029319 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clrq5\" (UniqueName: \"kubernetes.io/projected/fb68c844-6d7b-4f52-9a57-1ba8a4603449-kube-api-access-clrq5\") pod \"barbican-b26c-account-create-update-92x4v\" (UID: \"fb68c844-6d7b-4f52-9a57-1ba8a4603449\") " pod="openstack/barbican-b26c-account-create-update-92x4v" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.030260 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb68c844-6d7b-4f52-9a57-1ba8a4603449-operator-scripts\") pod \"barbican-b26c-account-create-update-92x4v\" (UID: \"fb68c844-6d7b-4f52-9a57-1ba8a4603449\") " pod="openstack/barbican-b26c-account-create-update-92x4v" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.030276 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qfblm" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.031657 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d14133a-bc13-40b8-9c84-cad4a04bb32e-operator-scripts\") pod \"barbican-db-create-hnwz5\" (UID: \"3d14133a-bc13-40b8-9c84-cad4a04bb32e\") " pod="openstack/barbican-db-create-hnwz5" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.036291 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-zxwkn"] Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.063393 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg5r9\" (UniqueName: \"kubernetes.io/projected/3d14133a-bc13-40b8-9c84-cad4a04bb32e-kube-api-access-tg5r9\") pod \"barbican-db-create-hnwz5\" (UID: \"3d14133a-bc13-40b8-9c84-cad4a04bb32e\") " pod="openstack/barbican-db-create-hnwz5" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.088443 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clrq5\" (UniqueName: \"kubernetes.io/projected/fb68c844-6d7b-4f52-9a57-1ba8a4603449-kube-api-access-clrq5\") pod \"barbican-b26c-account-create-update-92x4v\" (UID: \"fb68c844-6d7b-4f52-9a57-1ba8a4603449\") " pod="openstack/barbican-b26c-account-create-update-92x4v" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.103110 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-hnwz5" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.134139 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b26c-account-create-update-92x4v" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.134938 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5722ebb-3612-4608-9768-6d8a95d86efe-operator-scripts\") pod \"cinder-3040-account-create-update-mx8t8\" (UID: \"f5722ebb-3612-4608-9768-6d8a95d86efe\") " pod="openstack/cinder-3040-account-create-update-mx8t8" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.135990 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3baeace-c438-4e75-b503-a6b4ee66e19e-operator-scripts\") pod \"neutron-db-create-zxwkn\" (UID: \"b3baeace-c438-4e75-b503-a6b4ee66e19e\") " pod="openstack/neutron-db-create-zxwkn" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.136045 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qfs6\" (UniqueName: \"kubernetes.io/projected/b3baeace-c438-4e75-b503-a6b4ee66e19e-kube-api-access-5qfs6\") pod \"neutron-db-create-zxwkn\" (UID: \"b3baeace-c438-4e75-b503-a6b4ee66e19e\") " pod="openstack/neutron-db-create-zxwkn" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.136064 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5mgp\" (UniqueName: \"kubernetes.io/projected/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-kube-api-access-m5mgp\") pod \"keystone-db-sync-6z2v4\" (UID: \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\") " pod="openstack/keystone-db-sync-6z2v4" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.136106 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-config-data\") pod \"keystone-db-sync-6z2v4\" (UID: \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\") " pod="openstack/keystone-db-sync-6z2v4" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.136135 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zstf6\" (UniqueName: \"kubernetes.io/projected/f5722ebb-3612-4608-9768-6d8a95d86efe-kube-api-access-zstf6\") pod \"cinder-3040-account-create-update-mx8t8\" (UID: \"f5722ebb-3612-4608-9768-6d8a95d86efe\") " pod="openstack/cinder-3040-account-create-update-mx8t8" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.136177 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-combined-ca-bundle\") pod \"keystone-db-sync-6z2v4\" (UID: \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\") " pod="openstack/keystone-db-sync-6z2v4" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.135936 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5722ebb-3612-4608-9768-6d8a95d86efe-operator-scripts\") pod \"cinder-3040-account-create-update-mx8t8\" (UID: \"f5722ebb-3612-4608-9768-6d8a95d86efe\") " pod="openstack/cinder-3040-account-create-update-mx8t8" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.180306 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zstf6\" (UniqueName: \"kubernetes.io/projected/f5722ebb-3612-4608-9768-6d8a95d86efe-kube-api-access-zstf6\") pod \"cinder-3040-account-create-update-mx8t8\" (UID: \"f5722ebb-3612-4608-9768-6d8a95d86efe\") " pod="openstack/cinder-3040-account-create-update-mx8t8" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.207087 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3040-account-create-update-mx8t8" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.223037 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-05f0-account-create-update-nbjzz"] Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.224133 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-05f0-account-create-update-nbjzz" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.228083 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.246108 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3baeace-c438-4e75-b503-a6b4ee66e19e-operator-scripts\") pod \"neutron-db-create-zxwkn\" (UID: \"b3baeace-c438-4e75-b503-a6b4ee66e19e\") " pod="openstack/neutron-db-create-zxwkn" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.246157 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qfs6\" (UniqueName: \"kubernetes.io/projected/b3baeace-c438-4e75-b503-a6b4ee66e19e-kube-api-access-5qfs6\") pod \"neutron-db-create-zxwkn\" (UID: \"b3baeace-c438-4e75-b503-a6b4ee66e19e\") " pod="openstack/neutron-db-create-zxwkn" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.246176 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5mgp\" (UniqueName: \"kubernetes.io/projected/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-kube-api-access-m5mgp\") pod \"keystone-db-sync-6z2v4\" (UID: \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\") " pod="openstack/keystone-db-sync-6z2v4" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.246199 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-config-data\") pod \"keystone-db-sync-6z2v4\" (UID: \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\") " pod="openstack/keystone-db-sync-6z2v4" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.246232 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-combined-ca-bundle\") pod \"keystone-db-sync-6z2v4\" (UID: \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\") " pod="openstack/keystone-db-sync-6z2v4" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.248620 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3baeace-c438-4e75-b503-a6b4ee66e19e-operator-scripts\") pod \"neutron-db-create-zxwkn\" (UID: \"b3baeace-c438-4e75-b503-a6b4ee66e19e\") " pod="openstack/neutron-db-create-zxwkn" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.248809 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-05f0-account-create-update-nbjzz"] Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.258392 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-combined-ca-bundle\") pod \"keystone-db-sync-6z2v4\" (UID: \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\") " pod="openstack/keystone-db-sync-6z2v4" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.262006 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-config-data\") pod \"keystone-db-sync-6z2v4\" (UID: \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\") " pod="openstack/keystone-db-sync-6z2v4" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.276081 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-pm699" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.282295 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qfs6\" (UniqueName: \"kubernetes.io/projected/b3baeace-c438-4e75-b503-a6b4ee66e19e-kube-api-access-5qfs6\") pod \"neutron-db-create-zxwkn\" (UID: \"b3baeace-c438-4e75-b503-a6b4ee66e19e\") " pod="openstack/neutron-db-create-zxwkn" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.282927 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5mgp\" (UniqueName: \"kubernetes.io/projected/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-kube-api-access-m5mgp\") pod \"keystone-db-sync-6z2v4\" (UID: \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\") " pod="openstack/keystone-db-sync-6z2v4" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.284240 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6z2v4" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.334563 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zxwkn" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.363303 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p28xl\" (UniqueName: \"kubernetes.io/projected/a2e81ce0-545f-43bc-b2c2-bea1f931542d-kube-api-access-p28xl\") pod \"neutron-05f0-account-create-update-nbjzz\" (UID: \"a2e81ce0-545f-43bc-b2c2-bea1f931542d\") " pod="openstack/neutron-05f0-account-create-update-nbjzz" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.363373 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2e81ce0-545f-43bc-b2c2-bea1f931542d-operator-scripts\") pod \"neutron-05f0-account-create-update-nbjzz\" (UID: \"a2e81ce0-545f-43bc-b2c2-bea1f931542d\") " pod="openstack/neutron-05f0-account-create-update-nbjzz" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.465802 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p28xl\" (UniqueName: \"kubernetes.io/projected/a2e81ce0-545f-43bc-b2c2-bea1f931542d-kube-api-access-p28xl\") pod \"neutron-05f0-account-create-update-nbjzz\" (UID: \"a2e81ce0-545f-43bc-b2c2-bea1f931542d\") " pod="openstack/neutron-05f0-account-create-update-nbjzz" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.465881 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2e81ce0-545f-43bc-b2c2-bea1f931542d-operator-scripts\") pod \"neutron-05f0-account-create-update-nbjzz\" (UID: \"a2e81ce0-545f-43bc-b2c2-bea1f931542d\") " pod="openstack/neutron-05f0-account-create-update-nbjzz" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.466672 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2e81ce0-545f-43bc-b2c2-bea1f931542d-operator-scripts\") pod \"neutron-05f0-account-create-update-nbjzz\" (UID: \"a2e81ce0-545f-43bc-b2c2-bea1f931542d\") " pod="openstack/neutron-05f0-account-create-update-nbjzz" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.505659 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p28xl\" (UniqueName: \"kubernetes.io/projected/a2e81ce0-545f-43bc-b2c2-bea1f931542d-kube-api-access-p28xl\") pod \"neutron-05f0-account-create-update-nbjzz\" (UID: \"a2e81ce0-545f-43bc-b2c2-bea1f931542d\") " pod="openstack/neutron-05f0-account-create-update-nbjzz" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.567409 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.599084 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-05f0-account-create-update-nbjzz" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.670554 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-prometheus-metric-storage-rulefiles-0\") pod \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.670589 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-config\") pod \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.670773 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") pod \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.670821 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-thanos-prometheus-http-client-file\") pod \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.670847 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-web-config\") pod \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.671964 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-tls-assets\") pod \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.672020 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "c8bb2562-5d06-44d1-bbd0-8848d92f18d2" (UID: "c8bb2562-5d06-44d1-bbd0-8848d92f18d2"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.671995 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-config-out\") pod \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.672088 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfstp\" (UniqueName: \"kubernetes.io/projected/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-kube-api-access-dfstp\") pod \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\" (UID: \"c8bb2562-5d06-44d1-bbd0-8848d92f18d2\") " Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.672300 4884 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.679569 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-config" (OuterVolumeSpecName: "config") pod "c8bb2562-5d06-44d1-bbd0-8848d92f18d2" (UID: "c8bb2562-5d06-44d1-bbd0-8848d92f18d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.679686 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "c8bb2562-5d06-44d1-bbd0-8848d92f18d2" (UID: "c8bb2562-5d06-44d1-bbd0-8848d92f18d2"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.686504 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-kube-api-access-dfstp" (OuterVolumeSpecName: "kube-api-access-dfstp") pod "c8bb2562-5d06-44d1-bbd0-8848d92f18d2" (UID: "c8bb2562-5d06-44d1-bbd0-8848d92f18d2"). InnerVolumeSpecName "kube-api-access-dfstp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.686703 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-config-out" (OuterVolumeSpecName: "config-out") pod "c8bb2562-5d06-44d1-bbd0-8848d92f18d2" (UID: "c8bb2562-5d06-44d1-bbd0-8848d92f18d2"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.686863 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "c8bb2562-5d06-44d1-bbd0-8848d92f18d2" (UID: "c8bb2562-5d06-44d1-bbd0-8848d92f18d2"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.731557 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "c8bb2562-5d06-44d1-bbd0-8848d92f18d2" (UID: "c8bb2562-5d06-44d1-bbd0-8848d92f18d2"). InnerVolumeSpecName "pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.761617 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-web-config" (OuterVolumeSpecName: "web-config") pod "c8bb2562-5d06-44d1-bbd0-8848d92f18d2" (UID: "c8bb2562-5d06-44d1-bbd0-8848d92f18d2"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.773342 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mrh2r" event={"ID":"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e","Type":"ContainerStarted","Data":"20c36ae08413ad701e4df76588767073583a85dac1fbb1d94ea3e0188f80704b"} Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.783622 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-hnwz5"] Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.786927 4884 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.786960 4884 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-config-out\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.786970 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfstp\" (UniqueName: \"kubernetes.io/projected/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-kube-api-access-dfstp\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.786982 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.787006 4884 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") on node \"crc\" " Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.787016 4884 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.787025 4884 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/c8bb2562-5d06-44d1-bbd0-8848d92f18d2-web-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.787588 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-mrh2r" podStartSLOduration=3.1474294130000002 podStartE2EDuration="17.787574733s" podCreationTimestamp="2025-12-02 01:55:30 +0000 UTC" firstStartedPulling="2025-12-02 01:55:31.239361553 +0000 UTC m=+1147.915198477" lastFinishedPulling="2025-12-02 01:55:45.879506893 +0000 UTC m=+1162.555343797" observedRunningTime="2025-12-02 01:55:47.784566717 +0000 UTC m=+1164.460403601" watchObservedRunningTime="2025-12-02 01:55:47.787574733 +0000 UTC m=+1164.463411617" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.794017 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"c8bb2562-5d06-44d1-bbd0-8848d92f18d2","Type":"ContainerDied","Data":"812130cbe2c9e5d1331a385a71c447a151390e2b49670622b49eeabdaabdc691"} Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.794064 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.794068 4884 scope.go:117] "RemoveContainer" containerID="8a72e4291e35b4fb3bca6fa907132a6a3137ecf42f9d5a1e0c06bd5bf7d4c831" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.895360 4884 scope.go:117] "RemoveContainer" containerID="38a9fea55138df4195eab6c884265c3d5ff494f21607ba70544cff03fee07d66" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.897298 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.900401 4884 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.900537 4884 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05") on node "crc" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.911404 4884 reconciler_common.go:293] "Volume detached for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.927027 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.935709 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 02 01:55:47 crc kubenswrapper[4884]: E1202 01:55:47.936127 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerName="thanos-sidecar" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.936145 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerName="thanos-sidecar" Dec 02 01:55:47 crc kubenswrapper[4884]: E1202 01:55:47.936169 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerName="prometheus" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.936176 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerName="prometheus" Dec 02 01:55:47 crc kubenswrapper[4884]: E1202 01:55:47.936193 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerName="config-reloader" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.936200 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerName="config-reloader" Dec 02 01:55:47 crc kubenswrapper[4884]: E1202 01:55:47.936209 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerName="init-config-reloader" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.936216 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerName="init-config-reloader" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.937658 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerName="prometheus" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.937680 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerName="config-reloader" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.937692 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" containerName="thanos-sidecar" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.939176 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.942195 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.945947 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.946109 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.946299 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-mcslp" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.946550 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.946652 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.946739 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.951462 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qfblm"] Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.961116 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 02 01:55:47 crc kubenswrapper[4884]: I1202 01:55:47.987873 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3040-account-create-update-mx8t8"] Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.013064 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aa5b4806-95dc-4496-81cd-d66b3ad23c13-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.013094 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aa5b4806-95dc-4496-81cd-d66b3ad23c13-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.013114 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.013139 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aa5b4806-95dc-4496-81cd-d66b3ad23c13-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.013159 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.013221 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-config\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.013245 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.013262 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.013285 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.013303 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.013320 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5ngs\" (UniqueName: \"kubernetes.io/projected/aa5b4806-95dc-4496-81cd-d66b3ad23c13-kube-api-access-r5ngs\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.020061 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b26c-account-create-update-92x4v"] Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.025251 4884 scope.go:117] "RemoveContainer" containerID="05e715ded10f8d73e952217f675ad561ba62a82b52a6306b6dd3b7b674d8329e" Dec 02 01:55:48 crc kubenswrapper[4884]: W1202 01:55:48.071944 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb68c844_6d7b_4f52_9a57_1ba8a4603449.slice/crio-a1409e1e9353a42807ecf3303bfe8c06c76a19cc2db66bc73b72fb98fdebbea9 WatchSource:0}: Error finding container a1409e1e9353a42807ecf3303bfe8c06c76a19cc2db66bc73b72fb98fdebbea9: Status 404 returned error can't find the container with id a1409e1e9353a42807ecf3303bfe8c06c76a19cc2db66bc73b72fb98fdebbea9 Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.115284 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-config\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.116130 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.116167 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.116193 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.116210 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.116229 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5ngs\" (UniqueName: \"kubernetes.io/projected/aa5b4806-95dc-4496-81cd-d66b3ad23c13-kube-api-access-r5ngs\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.116271 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aa5b4806-95dc-4496-81cd-d66b3ad23c13-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.116290 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aa5b4806-95dc-4496-81cd-d66b3ad23c13-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.116309 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.116345 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aa5b4806-95dc-4496-81cd-d66b3ad23c13-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.116372 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.120136 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aa5b4806-95dc-4496-81cd-d66b3ad23c13-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.122021 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-config\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.122661 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.123510 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aa5b4806-95dc-4496-81cd-d66b3ad23c13-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.125436 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.128000 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aa5b4806-95dc-4496-81cd-d66b3ad23c13-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.128529 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.128934 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.129395 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.136469 4884 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.136498 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1350ca4960854133c6cd5b98efbfa142de46bc322b82e157adbca982350a030b/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.146020 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5ngs\" (UniqueName: \"kubernetes.io/projected/aa5b4806-95dc-4496-81cd-d66b3ad23c13-kube-api-access-r5ngs\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.177137 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-db-sync-pm699"] Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.187692 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-6z2v4"] Dec 02 01:55:48 crc kubenswrapper[4884]: W1202 01:55:48.219094 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8ccc2e4_81c2_434b_a9d4_62fb3e1a4443.slice/crio-503178a707798f60fd27f6d94b6968ab68e3ce2f233e3631653539f8048bcba8 WatchSource:0}: Error finding container 503178a707798f60fd27f6d94b6968ab68e3ce2f233e3631653539f8048bcba8: Status 404 returned error can't find the container with id 503178a707798f60fd27f6d94b6968ab68e3ce2f233e3631653539f8048bcba8 Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.238209 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-05f0-account-create-update-nbjzz"] Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.239229 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") pod \"prometheus-metric-storage-0\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.280165 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.331906 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-zxwkn"] Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.372624 4884 scope.go:117] "RemoveContainer" containerID="f632bbc70feae9bcab6e61183a7487126f6504963a2bb6247f36891b3ef48634" Dec 02 01:55:48 crc kubenswrapper[4884]: W1202 01:55:48.398075 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3baeace_c438_4e75_b503_a6b4ee66e19e.slice/crio-afe52fb7ae35a802abf32fda799fa35e35406dfd8132b2d24c16b1842e49a7fe WatchSource:0}: Error finding container afe52fb7ae35a802abf32fda799fa35e35406dfd8132b2d24c16b1842e49a7fe: Status 404 returned error can't find the container with id afe52fb7ae35a802abf32fda799fa35e35406dfd8132b2d24c16b1842e49a7fe Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.810569 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6z2v4" event={"ID":"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443","Type":"ContainerStarted","Data":"503178a707798f60fd27f6d94b6968ab68e3ce2f233e3631653539f8048bcba8"} Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.813503 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-pm699" event={"ID":"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8","Type":"ContainerStarted","Data":"f46a8eaec2f4213573dff68f66db7462342045321d98f1790820d49739ce6d94"} Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.830767 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-hnwz5" event={"ID":"3d14133a-bc13-40b8-9c84-cad4a04bb32e","Type":"ContainerStarted","Data":"5e012e78f633f504286f1a3f8d341b4f23db00e0ffd5bb95de81eb01118a7cfd"} Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.830805 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-hnwz5" event={"ID":"3d14133a-bc13-40b8-9c84-cad4a04bb32e","Type":"ContainerStarted","Data":"1cd77fedbc5fb3a7e3c1e0ceadcad93e971aa8cb7df417ea84c9eb94039b77e5"} Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.833782 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zxwkn" event={"ID":"b3baeace-c438-4e75-b503-a6b4ee66e19e","Type":"ContainerStarted","Data":"afe52fb7ae35a802abf32fda799fa35e35406dfd8132b2d24c16b1842e49a7fe"} Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.846414 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-05f0-account-create-update-nbjzz" event={"ID":"a2e81ce0-545f-43bc-b2c2-bea1f931542d","Type":"ContainerStarted","Data":"559d83a50df7830f16bd5fc38db352cb8066dcf9889466f1d0113ad7fd4d74b9"} Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.846459 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-05f0-account-create-update-nbjzz" event={"ID":"a2e81ce0-545f-43bc-b2c2-bea1f931542d","Type":"ContainerStarted","Data":"681b7294fd6e8cdad89f46e61c0f5f0518b35280a29bfff5052f7867165618f1"} Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.850182 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b26c-account-create-update-92x4v" event={"ID":"fb68c844-6d7b-4f52-9a57-1ba8a4603449","Type":"ContainerStarted","Data":"5f2b59ff1edac607729bca3d7f71ea5ca3e28819646a4e320756e5f5f9174152"} Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.850213 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b26c-account-create-update-92x4v" event={"ID":"fb68c844-6d7b-4f52-9a57-1ba8a4603449","Type":"ContainerStarted","Data":"a1409e1e9353a42807ecf3303bfe8c06c76a19cc2db66bc73b72fb98fdebbea9"} Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.852763 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-hnwz5" podStartSLOduration=2.852732944 podStartE2EDuration="2.852732944s" podCreationTimestamp="2025-12-02 01:55:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:55:48.845142413 +0000 UTC m=+1165.520979297" watchObservedRunningTime="2025-12-02 01:55:48.852732944 +0000 UTC m=+1165.528569828" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.853488 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3040-account-create-update-mx8t8" event={"ID":"f5722ebb-3612-4608-9768-6d8a95d86efe","Type":"ContainerStarted","Data":"ccf829018519eb030ae2b299409452e04846a62f68872a2a04271bb793567753"} Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.853518 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3040-account-create-update-mx8t8" event={"ID":"f5722ebb-3612-4608-9768-6d8a95d86efe","Type":"ContainerStarted","Data":"a5aba3cd631dc2b3f759d30a48ed134dd1da4399eea5f00b25ac7a28c711ae18"} Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.860284 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qfblm" event={"ID":"e1749692-f3f0-4849-bdf5-73ce5bfec72a","Type":"ContainerStarted","Data":"c7fa8f152081d6f4e0e751ec4e3fc7e0aa68e1ac3f86827e68929bc7a0b4adc9"} Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.860321 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qfblm" event={"ID":"e1749692-f3f0-4849-bdf5-73ce5bfec72a","Type":"ContainerStarted","Data":"0aa2ce8f4fd790fcd3d5c9c0e9e6d32e353337f52107e5c98b3553fbcb3f9d11"} Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.873591 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-05f0-account-create-update-nbjzz" podStartSLOduration=1.873570569 podStartE2EDuration="1.873570569s" podCreationTimestamp="2025-12-02 01:55:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:55:48.862047989 +0000 UTC m=+1165.537884873" watchObservedRunningTime="2025-12-02 01:55:48.873570569 +0000 UTC m=+1165.549407453" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.884853 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-3040-account-create-update-mx8t8" podStartSLOduration=2.884835232 podStartE2EDuration="2.884835232s" podCreationTimestamp="2025-12-02 01:55:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:55:48.877870347 +0000 UTC m=+1165.553707241" watchObservedRunningTime="2025-12-02 01:55:48.884835232 +0000 UTC m=+1165.560672116" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.898423 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-qfblm" podStartSLOduration=2.898402044 podStartE2EDuration="2.898402044s" podCreationTimestamp="2025-12-02 01:55:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:55:48.892230238 +0000 UTC m=+1165.568067122" watchObservedRunningTime="2025-12-02 01:55:48.898402044 +0000 UTC m=+1165.574238928" Dec 02 01:55:48 crc kubenswrapper[4884]: I1202 01:55:48.910338 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-b26c-account-create-update-92x4v" podStartSLOduration=2.910320043 podStartE2EDuration="2.910320043s" podCreationTimestamp="2025-12-02 01:55:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:55:48.905626425 +0000 UTC m=+1165.581463309" watchObservedRunningTime="2025-12-02 01:55:48.910320043 +0000 UTC m=+1165.586156927" Dec 02 01:55:49 crc kubenswrapper[4884]: I1202 01:55:49.018037 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 02 01:55:49 crc kubenswrapper[4884]: I1202 01:55:49.637727 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8bb2562-5d06-44d1-bbd0-8848d92f18d2" path="/var/lib/kubelet/pods/c8bb2562-5d06-44d1-bbd0-8848d92f18d2/volumes" Dec 02 01:55:49 crc kubenswrapper[4884]: I1202 01:55:49.868371 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zxwkn" event={"ID":"b3baeace-c438-4e75-b503-a6b4ee66e19e","Type":"ContainerStarted","Data":"92006ba5455afa5d492d493831b1d3ab7855aeeb503b8be1da12487ad1333afe"} Dec 02 01:55:49 crc kubenswrapper[4884]: I1202 01:55:49.870210 4884 generic.go:334] "Generic (PLEG): container finished" podID="a2e81ce0-545f-43bc-b2c2-bea1f931542d" containerID="559d83a50df7830f16bd5fc38db352cb8066dcf9889466f1d0113ad7fd4d74b9" exitCode=0 Dec 02 01:55:49 crc kubenswrapper[4884]: I1202 01:55:49.870251 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-05f0-account-create-update-nbjzz" event={"ID":"a2e81ce0-545f-43bc-b2c2-bea1f931542d","Type":"ContainerDied","Data":"559d83a50df7830f16bd5fc38db352cb8066dcf9889466f1d0113ad7fd4d74b9"} Dec 02 01:55:49 crc kubenswrapper[4884]: I1202 01:55:49.872064 4884 generic.go:334] "Generic (PLEG): container finished" podID="fb68c844-6d7b-4f52-9a57-1ba8a4603449" containerID="5f2b59ff1edac607729bca3d7f71ea5ca3e28819646a4e320756e5f5f9174152" exitCode=0 Dec 02 01:55:49 crc kubenswrapper[4884]: I1202 01:55:49.872116 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b26c-account-create-update-92x4v" event={"ID":"fb68c844-6d7b-4f52-9a57-1ba8a4603449","Type":"ContainerDied","Data":"5f2b59ff1edac607729bca3d7f71ea5ca3e28819646a4e320756e5f5f9174152"} Dec 02 01:55:49 crc kubenswrapper[4884]: I1202 01:55:49.873583 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"7605227ac5093bf15e557b7311797283006f54c319e7486aa601d9d78f5f4b66"} Dec 02 01:55:49 crc kubenswrapper[4884]: I1202 01:55:49.874879 4884 generic.go:334] "Generic (PLEG): container finished" podID="3d14133a-bc13-40b8-9c84-cad4a04bb32e" containerID="5e012e78f633f504286f1a3f8d341b4f23db00e0ffd5bb95de81eb01118a7cfd" exitCode=0 Dec 02 01:55:49 crc kubenswrapper[4884]: I1202 01:55:49.874914 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-hnwz5" event={"ID":"3d14133a-bc13-40b8-9c84-cad4a04bb32e","Type":"ContainerDied","Data":"5e012e78f633f504286f1a3f8d341b4f23db00e0ffd5bb95de81eb01118a7cfd"} Dec 02 01:55:49 crc kubenswrapper[4884]: I1202 01:55:49.876099 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aa5b4806-95dc-4496-81cd-d66b3ad23c13","Type":"ContainerStarted","Data":"7f609c966f2c9327164b9143efaf0eec277af4bb3a8fd991185c3f91876b8610"} Dec 02 01:55:49 crc kubenswrapper[4884]: I1202 01:55:49.877390 4884 generic.go:334] "Generic (PLEG): container finished" podID="f5722ebb-3612-4608-9768-6d8a95d86efe" containerID="ccf829018519eb030ae2b299409452e04846a62f68872a2a04271bb793567753" exitCode=0 Dec 02 01:55:49 crc kubenswrapper[4884]: I1202 01:55:49.877436 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3040-account-create-update-mx8t8" event={"ID":"f5722ebb-3612-4608-9768-6d8a95d86efe","Type":"ContainerDied","Data":"ccf829018519eb030ae2b299409452e04846a62f68872a2a04271bb793567753"} Dec 02 01:55:49 crc kubenswrapper[4884]: I1202 01:55:49.878737 4884 generic.go:334] "Generic (PLEG): container finished" podID="e1749692-f3f0-4849-bdf5-73ce5bfec72a" containerID="c7fa8f152081d6f4e0e751ec4e3fc7e0aa68e1ac3f86827e68929bc7a0b4adc9" exitCode=0 Dec 02 01:55:49 crc kubenswrapper[4884]: I1202 01:55:49.878783 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qfblm" event={"ID":"e1749692-f3f0-4849-bdf5-73ce5bfec72a","Type":"ContainerDied","Data":"c7fa8f152081d6f4e0e751ec4e3fc7e0aa68e1ac3f86827e68929bc7a0b4adc9"} Dec 02 01:55:50 crc kubenswrapper[4884]: I1202 01:55:50.892439 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"f71a30c04be4b377054fb0483f8f4c983d06f9bd24c3730f7cf4aec3b94ffd93"} Dec 02 01:55:50 crc kubenswrapper[4884]: I1202 01:55:50.908780 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-zxwkn" podStartSLOduration=4.908735636 podStartE2EDuration="4.908735636s" podCreationTimestamp="2025-12-02 01:55:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:55:50.907396643 +0000 UTC m=+1167.583233527" watchObservedRunningTime="2025-12-02 01:55:50.908735636 +0000 UTC m=+1167.584572520" Dec 02 01:55:51 crc kubenswrapper[4884]: I1202 01:55:51.900280 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aa5b4806-95dc-4496-81cd-d66b3ad23c13","Type":"ContainerStarted","Data":"5b5f08beb105d5a87ecc726ef6fdf90d4b742b063251973973341d0df9951e9f"} Dec 02 01:55:51 crc kubenswrapper[4884]: I1202 01:55:51.902737 4884 generic.go:334] "Generic (PLEG): container finished" podID="b3baeace-c438-4e75-b503-a6b4ee66e19e" containerID="92006ba5455afa5d492d493831b1d3ab7855aeeb503b8be1da12487ad1333afe" exitCode=0 Dec 02 01:55:51 crc kubenswrapper[4884]: I1202 01:55:51.902775 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zxwkn" event={"ID":"b3baeace-c438-4e75-b503-a6b4ee66e19e","Type":"ContainerDied","Data":"92006ba5455afa5d492d493831b1d3ab7855aeeb503b8be1da12487ad1333afe"} Dec 02 01:55:55 crc kubenswrapper[4884]: I1202 01:55:55.944853 4884 generic.go:334] "Generic (PLEG): container finished" podID="83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e" containerID="20c36ae08413ad701e4df76588767073583a85dac1fbb1d94ea3e0188f80704b" exitCode=0 Dec 02 01:55:55 crc kubenswrapper[4884]: I1202 01:55:55.944932 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mrh2r" event={"ID":"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e","Type":"ContainerDied","Data":"20c36ae08413ad701e4df76588767073583a85dac1fbb1d94ea3e0188f80704b"} Dec 02 01:55:57 crc kubenswrapper[4884]: I1202 01:55:57.904244 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-05f0-account-create-update-nbjzz" Dec 02 01:55:57 crc kubenswrapper[4884]: I1202 01:55:57.914286 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b26c-account-create-update-92x4v" Dec 02 01:55:57 crc kubenswrapper[4884]: I1202 01:55:57.949046 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zxwkn" Dec 02 01:55:57 crc kubenswrapper[4884]: I1202 01:55:57.954413 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-hnwz5" Dec 02 01:55:57 crc kubenswrapper[4884]: I1202 01:55:57.971336 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qfblm" Dec 02 01:55:57 crc kubenswrapper[4884]: I1202 01:55:57.991639 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:57 crc kubenswrapper[4884]: I1202 01:55:57.991936 4884 generic.go:334] "Generic (PLEG): container finished" podID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerID="5b5f08beb105d5a87ecc726ef6fdf90d4b742b063251973973341d0df9951e9f" exitCode=0 Dec 02 01:55:57 crc kubenswrapper[4884]: I1202 01:55:57.992065 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aa5b4806-95dc-4496-81cd-d66b3ad23c13","Type":"ContainerDied","Data":"5b5f08beb105d5a87ecc726ef6fdf90d4b742b063251973973341d0df9951e9f"} Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.000583 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3040-account-create-update-mx8t8" event={"ID":"f5722ebb-3612-4608-9768-6d8a95d86efe","Type":"ContainerDied","Data":"a5aba3cd631dc2b3f759d30a48ed134dd1da4399eea5f00b25ac7a28c711ae18"} Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.000624 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5aba3cd631dc2b3f759d30a48ed134dd1da4399eea5f00b25ac7a28c711ae18" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.002456 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3040-account-create-update-mx8t8" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.014182 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zxwkn" event={"ID":"b3baeace-c438-4e75-b503-a6b4ee66e19e","Type":"ContainerDied","Data":"afe52fb7ae35a802abf32fda799fa35e35406dfd8132b2d24c16b1842e49a7fe"} Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.014227 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afe52fb7ae35a802abf32fda799fa35e35406dfd8132b2d24c16b1842e49a7fe" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.014312 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zxwkn" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.022528 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qfblm" event={"ID":"e1749692-f3f0-4849-bdf5-73ce5bfec72a","Type":"ContainerDied","Data":"0aa2ce8f4fd790fcd3d5c9c0e9e6d32e353337f52107e5c98b3553fbcb3f9d11"} Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.022580 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0aa2ce8f4fd790fcd3d5c9c0e9e6d32e353337f52107e5c98b3553fbcb3f9d11" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.022658 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qfblm" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.026164 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mrh2r" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.026385 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mrh2r" event={"ID":"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e","Type":"ContainerDied","Data":"2d3e14c4cf3fd1403b1db225c5fc85b2a852681b84a51dd4233f9661ca8c3cb4"} Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.026414 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d3e14c4cf3fd1403b1db225c5fc85b2a852681b84a51dd4233f9661ca8c3cb4" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.028091 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-05f0-account-create-update-nbjzz" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.028474 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-05f0-account-create-update-nbjzz" event={"ID":"a2e81ce0-545f-43bc-b2c2-bea1f931542d","Type":"ContainerDied","Data":"681b7294fd6e8cdad89f46e61c0f5f0518b35280a29bfff5052f7867165618f1"} Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.028502 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="681b7294fd6e8cdad89f46e61c0f5f0518b35280a29bfff5052f7867165618f1" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.033086 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b26c-account-create-update-92x4v" event={"ID":"fb68c844-6d7b-4f52-9a57-1ba8a4603449","Type":"ContainerDied","Data":"a1409e1e9353a42807ecf3303bfe8c06c76a19cc2db66bc73b72fb98fdebbea9"} Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.033113 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1409e1e9353a42807ecf3303bfe8c06c76a19cc2db66bc73b72fb98fdebbea9" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.033159 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b26c-account-create-update-92x4v" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.041784 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-hnwz5" event={"ID":"3d14133a-bc13-40b8-9c84-cad4a04bb32e","Type":"ContainerDied","Data":"1cd77fedbc5fb3a7e3c1e0ceadcad93e971aa8cb7df417ea84c9eb94039b77e5"} Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.041823 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cd77fedbc5fb3a7e3c1e0ceadcad93e971aa8cb7df417ea84c9eb94039b77e5" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.041851 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-hnwz5" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.054393 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2e81ce0-545f-43bc-b2c2-bea1f931542d-operator-scripts\") pod \"a2e81ce0-545f-43bc-b2c2-bea1f931542d\" (UID: \"a2e81ce0-545f-43bc-b2c2-bea1f931542d\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.054484 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb68c844-6d7b-4f52-9a57-1ba8a4603449-operator-scripts\") pod \"fb68c844-6d7b-4f52-9a57-1ba8a4603449\" (UID: \"fb68c844-6d7b-4f52-9a57-1ba8a4603449\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.054525 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d14133a-bc13-40b8-9c84-cad4a04bb32e-operator-scripts\") pod \"3d14133a-bc13-40b8-9c84-cad4a04bb32e\" (UID: \"3d14133a-bc13-40b8-9c84-cad4a04bb32e\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.054600 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tg5r9\" (UniqueName: \"kubernetes.io/projected/3d14133a-bc13-40b8-9c84-cad4a04bb32e-kube-api-access-tg5r9\") pod \"3d14133a-bc13-40b8-9c84-cad4a04bb32e\" (UID: \"3d14133a-bc13-40b8-9c84-cad4a04bb32e\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.054628 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3baeace-c438-4e75-b503-a6b4ee66e19e-operator-scripts\") pod \"b3baeace-c438-4e75-b503-a6b4ee66e19e\" (UID: \"b3baeace-c438-4e75-b503-a6b4ee66e19e\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.054801 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clrq5\" (UniqueName: \"kubernetes.io/projected/fb68c844-6d7b-4f52-9a57-1ba8a4603449-kube-api-access-clrq5\") pod \"fb68c844-6d7b-4f52-9a57-1ba8a4603449\" (UID: \"fb68c844-6d7b-4f52-9a57-1ba8a4603449\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.054859 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p28xl\" (UniqueName: \"kubernetes.io/projected/a2e81ce0-545f-43bc-b2c2-bea1f931542d-kube-api-access-p28xl\") pod \"a2e81ce0-545f-43bc-b2c2-bea1f931542d\" (UID: \"a2e81ce0-545f-43bc-b2c2-bea1f931542d\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.054960 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qfs6\" (UniqueName: \"kubernetes.io/projected/b3baeace-c438-4e75-b503-a6b4ee66e19e-kube-api-access-5qfs6\") pod \"b3baeace-c438-4e75-b503-a6b4ee66e19e\" (UID: \"b3baeace-c438-4e75-b503-a6b4ee66e19e\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.055242 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e81ce0-545f-43bc-b2c2-bea1f931542d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a2e81ce0-545f-43bc-b2c2-bea1f931542d" (UID: "a2e81ce0-545f-43bc-b2c2-bea1f931542d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.055303 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb68c844-6d7b-4f52-9a57-1ba8a4603449-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fb68c844-6d7b-4f52-9a57-1ba8a4603449" (UID: "fb68c844-6d7b-4f52-9a57-1ba8a4603449"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.055860 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d14133a-bc13-40b8-9c84-cad4a04bb32e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3d14133a-bc13-40b8-9c84-cad4a04bb32e" (UID: "3d14133a-bc13-40b8-9c84-cad4a04bb32e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.056109 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2e81ce0-545f-43bc-b2c2-bea1f931542d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.056121 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb68c844-6d7b-4f52-9a57-1ba8a4603449-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.056131 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3d14133a-bc13-40b8-9c84-cad4a04bb32e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.056502 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3baeace-c438-4e75-b503-a6b4ee66e19e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b3baeace-c438-4e75-b503-a6b4ee66e19e" (UID: "b3baeace-c438-4e75-b503-a6b4ee66e19e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.060168 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb68c844-6d7b-4f52-9a57-1ba8a4603449-kube-api-access-clrq5" (OuterVolumeSpecName: "kube-api-access-clrq5") pod "fb68c844-6d7b-4f52-9a57-1ba8a4603449" (UID: "fb68c844-6d7b-4f52-9a57-1ba8a4603449"). InnerVolumeSpecName "kube-api-access-clrq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.060251 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3baeace-c438-4e75-b503-a6b4ee66e19e-kube-api-access-5qfs6" (OuterVolumeSpecName: "kube-api-access-5qfs6") pod "b3baeace-c438-4e75-b503-a6b4ee66e19e" (UID: "b3baeace-c438-4e75-b503-a6b4ee66e19e"). InnerVolumeSpecName "kube-api-access-5qfs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.062511 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d14133a-bc13-40b8-9c84-cad4a04bb32e-kube-api-access-tg5r9" (OuterVolumeSpecName: "kube-api-access-tg5r9") pod "3d14133a-bc13-40b8-9c84-cad4a04bb32e" (UID: "3d14133a-bc13-40b8-9c84-cad4a04bb32e"). InnerVolumeSpecName "kube-api-access-tg5r9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.063663 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2e81ce0-545f-43bc-b2c2-bea1f931542d-kube-api-access-p28xl" (OuterVolumeSpecName: "kube-api-access-p28xl") pod "a2e81ce0-545f-43bc-b2c2-bea1f931542d" (UID: "a2e81ce0-545f-43bc-b2c2-bea1f931542d"). InnerVolumeSpecName "kube-api-access-p28xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.157540 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wrfk\" (UniqueName: \"kubernetes.io/projected/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-kube-api-access-9wrfk\") pod \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.157580 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5p4fb\" (UniqueName: \"kubernetes.io/projected/e1749692-f3f0-4849-bdf5-73ce5bfec72a-kube-api-access-5p4fb\") pod \"e1749692-f3f0-4849-bdf5-73ce5bfec72a\" (UID: \"e1749692-f3f0-4849-bdf5-73ce5bfec72a\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.157649 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1749692-f3f0-4849-bdf5-73ce5bfec72a-operator-scripts\") pod \"e1749692-f3f0-4849-bdf5-73ce5bfec72a\" (UID: \"e1749692-f3f0-4849-bdf5-73ce5bfec72a\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.157666 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5722ebb-3612-4608-9768-6d8a95d86efe-operator-scripts\") pod \"f5722ebb-3612-4608-9768-6d8a95d86efe\" (UID: \"f5722ebb-3612-4608-9768-6d8a95d86efe\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.157688 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-combined-ca-bundle\") pod \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.157857 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zstf6\" (UniqueName: \"kubernetes.io/projected/f5722ebb-3612-4608-9768-6d8a95d86efe-kube-api-access-zstf6\") pod \"f5722ebb-3612-4608-9768-6d8a95d86efe\" (UID: \"f5722ebb-3612-4608-9768-6d8a95d86efe\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.157884 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-config-data\") pod \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.157901 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-db-sync-config-data\") pod \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\" (UID: \"83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e\") " Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.158299 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qfs6\" (UniqueName: \"kubernetes.io/projected/b3baeace-c438-4e75-b503-a6b4ee66e19e-kube-api-access-5qfs6\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.158311 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tg5r9\" (UniqueName: \"kubernetes.io/projected/3d14133a-bc13-40b8-9c84-cad4a04bb32e-kube-api-access-tg5r9\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.158320 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3baeace-c438-4e75-b503-a6b4ee66e19e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.158329 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clrq5\" (UniqueName: \"kubernetes.io/projected/fb68c844-6d7b-4f52-9a57-1ba8a4603449-kube-api-access-clrq5\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.158337 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p28xl\" (UniqueName: \"kubernetes.io/projected/a2e81ce0-545f-43bc-b2c2-bea1f931542d-kube-api-access-p28xl\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.160603 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1749692-f3f0-4849-bdf5-73ce5bfec72a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e1749692-f3f0-4849-bdf5-73ce5bfec72a" (UID: "e1749692-f3f0-4849-bdf5-73ce5bfec72a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.160900 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5722ebb-3612-4608-9768-6d8a95d86efe-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f5722ebb-3612-4608-9768-6d8a95d86efe" (UID: "f5722ebb-3612-4608-9768-6d8a95d86efe"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.162963 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e" (UID: "83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.163072 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-kube-api-access-9wrfk" (OuterVolumeSpecName: "kube-api-access-9wrfk") pod "83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e" (UID: "83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e"). InnerVolumeSpecName "kube-api-access-9wrfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.163490 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1749692-f3f0-4849-bdf5-73ce5bfec72a-kube-api-access-5p4fb" (OuterVolumeSpecName: "kube-api-access-5p4fb") pod "e1749692-f3f0-4849-bdf5-73ce5bfec72a" (UID: "e1749692-f3f0-4849-bdf5-73ce5bfec72a"). InnerVolumeSpecName "kube-api-access-5p4fb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.163519 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5722ebb-3612-4608-9768-6d8a95d86efe-kube-api-access-zstf6" (OuterVolumeSpecName: "kube-api-access-zstf6") pod "f5722ebb-3612-4608-9768-6d8a95d86efe" (UID: "f5722ebb-3612-4608-9768-6d8a95d86efe"). InnerVolumeSpecName "kube-api-access-zstf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.181801 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e" (UID: "83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.204528 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-config-data" (OuterVolumeSpecName: "config-data") pod "83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e" (UID: "83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.260871 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zstf6\" (UniqueName: \"kubernetes.io/projected/f5722ebb-3612-4608-9768-6d8a95d86efe-kube-api-access-zstf6\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.261128 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.261140 4884 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.261151 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wrfk\" (UniqueName: \"kubernetes.io/projected/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-kube-api-access-9wrfk\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.261163 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5p4fb\" (UniqueName: \"kubernetes.io/projected/e1749692-f3f0-4849-bdf5-73ce5bfec72a-kube-api-access-5p4fb\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.261171 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1749692-f3f0-4849-bdf5-73ce5bfec72a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.261180 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5722ebb-3612-4608-9768-6d8a95d86efe-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:58 crc kubenswrapper[4884]: I1202 01:55:58.261187 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.051026 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"c27cae3424a770818dc255185b93b432237c6066cabda73490f9ef02a0b1a886"} Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.051066 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3040-account-create-update-mx8t8" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.414545 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-mltjh"] Dec 02 01:55:59 crc kubenswrapper[4884]: E1202 01:55:59.416975 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2e81ce0-545f-43bc-b2c2-bea1f931542d" containerName="mariadb-account-create-update" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.417011 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2e81ce0-545f-43bc-b2c2-bea1f931542d" containerName="mariadb-account-create-update" Dec 02 01:55:59 crc kubenswrapper[4884]: E1202 01:55:59.417050 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5722ebb-3612-4608-9768-6d8a95d86efe" containerName="mariadb-account-create-update" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.417060 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5722ebb-3612-4608-9768-6d8a95d86efe" containerName="mariadb-account-create-update" Dec 02 01:55:59 crc kubenswrapper[4884]: E1202 01:55:59.417087 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3baeace-c438-4e75-b503-a6b4ee66e19e" containerName="mariadb-database-create" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.417097 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3baeace-c438-4e75-b503-a6b4ee66e19e" containerName="mariadb-database-create" Dec 02 01:55:59 crc kubenswrapper[4884]: E1202 01:55:59.417113 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e" containerName="glance-db-sync" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.417121 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e" containerName="glance-db-sync" Dec 02 01:55:59 crc kubenswrapper[4884]: E1202 01:55:59.417130 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1749692-f3f0-4849-bdf5-73ce5bfec72a" containerName="mariadb-database-create" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.417137 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1749692-f3f0-4849-bdf5-73ce5bfec72a" containerName="mariadb-database-create" Dec 02 01:55:59 crc kubenswrapper[4884]: E1202 01:55:59.417150 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb68c844-6d7b-4f52-9a57-1ba8a4603449" containerName="mariadb-account-create-update" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.417158 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb68c844-6d7b-4f52-9a57-1ba8a4603449" containerName="mariadb-account-create-update" Dec 02 01:55:59 crc kubenswrapper[4884]: E1202 01:55:59.417173 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d14133a-bc13-40b8-9c84-cad4a04bb32e" containerName="mariadb-database-create" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.417180 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d14133a-bc13-40b8-9c84-cad4a04bb32e" containerName="mariadb-database-create" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.417434 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d14133a-bc13-40b8-9c84-cad4a04bb32e" containerName="mariadb-database-create" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.417453 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5722ebb-3612-4608-9768-6d8a95d86efe" containerName="mariadb-account-create-update" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.417464 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2e81ce0-545f-43bc-b2c2-bea1f931542d" containerName="mariadb-account-create-update" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.417486 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb68c844-6d7b-4f52-9a57-1ba8a4603449" containerName="mariadb-account-create-update" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.417501 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1749692-f3f0-4849-bdf5-73ce5bfec72a" containerName="mariadb-database-create" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.417511 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e" containerName="glance-db-sync" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.417523 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3baeace-c438-4e75-b503-a6b4ee66e19e" containerName="mariadb-database-create" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.423098 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.434920 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-mltjh"] Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.590127 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-mltjh\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.590208 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-mltjh\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.590259 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-mltjh\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.590295 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpwq9\" (UniqueName: \"kubernetes.io/projected/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-kube-api-access-zpwq9\") pod \"dnsmasq-dns-5b946c75cc-mltjh\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.590316 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-config\") pod \"dnsmasq-dns-5b946c75cc-mltjh\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.692139 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpwq9\" (UniqueName: \"kubernetes.io/projected/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-kube-api-access-zpwq9\") pod \"dnsmasq-dns-5b946c75cc-mltjh\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.692473 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-config\") pod \"dnsmasq-dns-5b946c75cc-mltjh\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.692553 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-mltjh\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.692606 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-mltjh\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.692646 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-mltjh\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.693403 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-config\") pod \"dnsmasq-dns-5b946c75cc-mltjh\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.693430 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-mltjh\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.693454 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-mltjh\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.693543 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-mltjh\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.709473 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpwq9\" (UniqueName: \"kubernetes.io/projected/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-kube-api-access-zpwq9\") pod \"dnsmasq-dns-5b946c75cc-mltjh\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:55:59 crc kubenswrapper[4884]: I1202 01:55:59.748268 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:56:02 crc kubenswrapper[4884]: E1202 01:56:02.750363 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.5:5001/podified-epoxy-centos9/openstack-watcher-api:watcher_latest" Dec 02 01:56:02 crc kubenswrapper[4884]: E1202 01:56:02.751070 4884 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.5:5001/podified-epoxy-centos9/openstack-watcher-api:watcher_latest" Dec 02 01:56:02 crc kubenswrapper[4884]: E1202 01:56:02.751350 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:watcher-db-sync,Image:38.102.83.5:5001/podified-epoxy-centos9/openstack-watcher-api:watcher_latest,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/watcher/watcher.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:watcher-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lh94v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-db-sync-pm699_openstack(5ba1201c-49ba-4c31-bf2a-8069ea8a16c8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 01:56:02 crc kubenswrapper[4884]: E1202 01:56:02.754836 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/watcher-db-sync-pm699" podUID="5ba1201c-49ba-4c31-bf2a-8069ea8a16c8" Dec 02 01:56:03 crc kubenswrapper[4884]: I1202 01:56:03.104614 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"f2b6f2c806f82db6b820847e67ddd9361879262c98d8bf32a468f70adaef5528"} Dec 02 01:56:03 crc kubenswrapper[4884]: I1202 01:56:03.108446 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6z2v4" event={"ID":"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443","Type":"ContainerStarted","Data":"f50852b53c911e52af5cc0b9e21d2f0baa40e2bcd63cc95a9dcf9f9f9800b041"} Dec 02 01:56:03 crc kubenswrapper[4884]: I1202 01:56:03.110807 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aa5b4806-95dc-4496-81cd-d66b3ad23c13","Type":"ContainerStarted","Data":"ff2a72f07e42090c0d6e3254cf3b17dd9ef3b8f1b498858eb94632c5a0884931"} Dec 02 01:56:03 crc kubenswrapper[4884]: E1202 01:56:03.113173 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"watcher-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.5:5001/podified-epoxy-centos9/openstack-watcher-api:watcher_latest\\\"\"" pod="openstack/watcher-db-sync-pm699" podUID="5ba1201c-49ba-4c31-bf2a-8069ea8a16c8" Dec 02 01:56:03 crc kubenswrapper[4884]: I1202 01:56:03.129905 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-6z2v4" podStartSLOduration=2.652441682 podStartE2EDuration="17.129887342s" podCreationTimestamp="2025-12-02 01:55:46 +0000 UTC" firstStartedPulling="2025-12-02 01:55:48.244950501 +0000 UTC m=+1164.920787385" lastFinishedPulling="2025-12-02 01:56:02.722396161 +0000 UTC m=+1179.398233045" observedRunningTime="2025-12-02 01:56:03.125288788 +0000 UTC m=+1179.801125692" watchObservedRunningTime="2025-12-02 01:56:03.129887342 +0000 UTC m=+1179.805724226" Dec 02 01:56:03 crc kubenswrapper[4884]: I1202 01:56:03.225199 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-mltjh"] Dec 02 01:56:03 crc kubenswrapper[4884]: W1202 01:56:03.229287 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc347dd5_89f1_4f68_8d3b_c8fedaaa3c21.slice/crio-efb1a6f5eb68f020ddcdf4797d49066fbe59f9c0819e72fc2c13549717bc49d1 WatchSource:0}: Error finding container efb1a6f5eb68f020ddcdf4797d49066fbe59f9c0819e72fc2c13549717bc49d1: Status 404 returned error can't find the container with id efb1a6f5eb68f020ddcdf4797d49066fbe59f9c0819e72fc2c13549717bc49d1 Dec 02 01:56:04 crc kubenswrapper[4884]: I1202 01:56:04.121437 4884 generic.go:334] "Generic (PLEG): container finished" podID="bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21" containerID="b216246e9d169813390faa7f6252303d40feb3b5bb9be26c0fbd7371d103fb20" exitCode=0 Dec 02 01:56:04 crc kubenswrapper[4884]: I1202 01:56:04.121481 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" event={"ID":"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21","Type":"ContainerDied","Data":"b216246e9d169813390faa7f6252303d40feb3b5bb9be26c0fbd7371d103fb20"} Dec 02 01:56:04 crc kubenswrapper[4884]: I1202 01:56:04.121864 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" event={"ID":"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21","Type":"ContainerStarted","Data":"efb1a6f5eb68f020ddcdf4797d49066fbe59f9c0819e72fc2c13549717bc49d1"} Dec 02 01:56:05 crc kubenswrapper[4884]: I1202 01:56:05.135067 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" event={"ID":"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21","Type":"ContainerStarted","Data":"cf483adfbc1ccdf1397ec42465607bf68aee7b475be6035b2d90e4bbb3f49416"} Dec 02 01:56:05 crc kubenswrapper[4884]: I1202 01:56:05.135403 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:56:06 crc kubenswrapper[4884]: I1202 01:56:06.153721 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"4ed3832d11cd6e2d1ddf23075f9c5e06770e108c3cf60903ce1b9aa46f69c1b6"} Dec 02 01:56:07 crc kubenswrapper[4884]: I1202 01:56:07.165705 4884 generic.go:334] "Generic (PLEG): container finished" podID="a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443" containerID="f50852b53c911e52af5cc0b9e21d2f0baa40e2bcd63cc95a9dcf9f9f9800b041" exitCode=0 Dec 02 01:56:07 crc kubenswrapper[4884]: I1202 01:56:07.165830 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6z2v4" event={"ID":"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443","Type":"ContainerDied","Data":"f50852b53c911e52af5cc0b9e21d2f0baa40e2bcd63cc95a9dcf9f9f9800b041"} Dec 02 01:56:07 crc kubenswrapper[4884]: I1202 01:56:07.169618 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aa5b4806-95dc-4496-81cd-d66b3ad23c13","Type":"ContainerStarted","Data":"6103d609ead7e74a2f2a809ab023c79ca78a2835c48a603c317ecbf6f95d38bd"} Dec 02 01:56:07 crc kubenswrapper[4884]: I1202 01:56:07.169700 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aa5b4806-95dc-4496-81cd-d66b3ad23c13","Type":"ContainerStarted","Data":"22cae3ab5f411ea5f30ea9e8ad3ff278d413825ae977a8ee143fd766e1cf651a"} Dec 02 01:56:07 crc kubenswrapper[4884]: I1202 01:56:07.173676 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"30a8632e76553e5dd4abca9ec213d4f642751cfca95b70a072b73c44a636bc1c"} Dec 02 01:56:07 crc kubenswrapper[4884]: I1202 01:56:07.173715 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"24d2d4810ac3e803d025f2cc39f835b358dd4bb2abd2bd50c3bcbd8f838773ac"} Dec 02 01:56:07 crc kubenswrapper[4884]: I1202 01:56:07.173726 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"c5460c30e14e09ed38d8a641aef9beef7a466e30c19483279900cb53e4cc196b"} Dec 02 01:56:07 crc kubenswrapper[4884]: I1202 01:56:07.198649 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" podStartSLOduration=8.198619995 podStartE2EDuration="8.198619995s" podCreationTimestamp="2025-12-02 01:55:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:05.163868878 +0000 UTC m=+1181.839705832" watchObservedRunningTime="2025-12-02 01:56:07.198619995 +0000 UTC m=+1183.874456919" Dec 02 01:56:07 crc kubenswrapper[4884]: I1202 01:56:07.236638 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=20.236616953 podStartE2EDuration="20.236616953s" podCreationTimestamp="2025-12-02 01:55:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:07.227149589 +0000 UTC m=+1183.902986473" watchObservedRunningTime="2025-12-02 01:56:07.236616953 +0000 UTC m=+1183.912453847" Dec 02 01:56:08 crc kubenswrapper[4884]: I1202 01:56:08.280982 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 02 01:56:08 crc kubenswrapper[4884]: I1202 01:56:08.648595 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6z2v4" Dec 02 01:56:08 crc kubenswrapper[4884]: I1202 01:56:08.773380 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5mgp\" (UniqueName: \"kubernetes.io/projected/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-kube-api-access-m5mgp\") pod \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\" (UID: \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\") " Dec 02 01:56:08 crc kubenswrapper[4884]: I1202 01:56:08.773749 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-combined-ca-bundle\") pod \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\" (UID: \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\") " Dec 02 01:56:08 crc kubenswrapper[4884]: I1202 01:56:08.773851 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-config-data\") pod \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\" (UID: \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\") " Dec 02 01:56:08 crc kubenswrapper[4884]: I1202 01:56:08.780989 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-kube-api-access-m5mgp" (OuterVolumeSpecName: "kube-api-access-m5mgp") pod "a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443" (UID: "a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443"). InnerVolumeSpecName "kube-api-access-m5mgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:08 crc kubenswrapper[4884]: I1202 01:56:08.829066 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443" (UID: "a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:08 crc kubenswrapper[4884]: I1202 01:56:08.878133 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-config-data" (OuterVolumeSpecName: "config-data") pod "a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443" (UID: "a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:08 crc kubenswrapper[4884]: I1202 01:56:08.878306 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-config-data\") pod \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\" (UID: \"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443\") " Dec 02 01:56:08 crc kubenswrapper[4884]: W1202 01:56:08.878693 4884 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443/volumes/kubernetes.io~secret/config-data Dec 02 01:56:08 crc kubenswrapper[4884]: I1202 01:56:08.878738 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-config-data" (OuterVolumeSpecName: "config-data") pod "a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443" (UID: "a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:08 crc kubenswrapper[4884]: I1202 01:56:08.879305 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:08 crc kubenswrapper[4884]: I1202 01:56:08.879347 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5mgp\" (UniqueName: \"kubernetes.io/projected/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-kube-api-access-m5mgp\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:08 crc kubenswrapper[4884]: I1202 01:56:08.879370 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.206501 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"55711d0ae8501fa53aac4cf745ff40e25f554c05b33935e4933dd29096d43e1c"} Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.206770 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"509cd38add1a0d7ce0a2216b01663f3d38621d1fd3cba7741d688dda996274d0"} Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.206869 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"752d466950de96385b6f85ef2bcfbb961e844ad78b50c4ad91b2fbaac1aba4e3"} Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.208979 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-6z2v4" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.210001 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-6z2v4" event={"ID":"a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443","Type":"ContainerDied","Data":"503178a707798f60fd27f6d94b6968ab68e3ce2f233e3631653539f8048bcba8"} Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.210152 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="503178a707798f60fd27f6d94b6968ab68e3ce2f233e3631653539f8048bcba8" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.481452 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-mltjh"] Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.481920 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" podUID="bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21" containerName="dnsmasq-dns" containerID="cri-o://cf483adfbc1ccdf1397ec42465607bf68aee7b475be6035b2d90e4bbb3f49416" gracePeriod=10 Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.483889 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.504396 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-4cqgj"] Dec 02 01:56:09 crc kubenswrapper[4884]: E1202 01:56:09.504759 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443" containerName="keystone-db-sync" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.504776 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443" containerName="keystone-db-sync" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.504958 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443" containerName="keystone-db-sync" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.505860 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.512925 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.513043 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.513270 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.513573 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.513805 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-42pg8" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.530002 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4cqgj"] Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.546960 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-784f69c749-cxssm"] Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.548264 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.595659 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcghm\" (UniqueName: \"kubernetes.io/projected/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-kube-api-access-gcghm\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.595839 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-credential-keys\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.595873 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-scripts\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.595891 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-config-data\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.595922 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-fernet-keys\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.595942 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-combined-ca-bundle\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.699568 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-ovsdbserver-sb\") pod \"dnsmasq-dns-784f69c749-cxssm\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.699638 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcghm\" (UniqueName: \"kubernetes.io/projected/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-kube-api-access-gcghm\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.699666 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-config\") pod \"dnsmasq-dns-784f69c749-cxssm\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.699696 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-ovsdbserver-nb\") pod \"dnsmasq-dns-784f69c749-cxssm\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.699720 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2cpx\" (UniqueName: \"kubernetes.io/projected/b29d1a05-edbf-4146-8c0f-21e846bad95f-kube-api-access-p2cpx\") pod \"dnsmasq-dns-784f69c749-cxssm\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.699780 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-credential-keys\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.699810 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-scripts\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.699829 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-config-data\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.699861 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-fernet-keys\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.699884 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-combined-ca-bundle\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.699909 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-dns-svc\") pod \"dnsmasq-dns-784f69c749-cxssm\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.723481 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-config-data\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.730390 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-credential-keys\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.737489 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcghm\" (UniqueName: \"kubernetes.io/projected/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-kube-api-access-gcghm\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.738055 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-fernet-keys\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.743074 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-scripts\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.753231 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" podUID="bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.141:5353: connect: connection refused" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.755362 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-combined-ca-bundle\") pod \"keystone-bootstrap-4cqgj\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.801766 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-config\") pod \"dnsmasq-dns-784f69c749-cxssm\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.801830 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-ovsdbserver-nb\") pod \"dnsmasq-dns-784f69c749-cxssm\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.801890 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2cpx\" (UniqueName: \"kubernetes.io/projected/b29d1a05-edbf-4146-8c0f-21e846bad95f-kube-api-access-p2cpx\") pod \"dnsmasq-dns-784f69c749-cxssm\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.802002 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-dns-svc\") pod \"dnsmasq-dns-784f69c749-cxssm\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.802041 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-ovsdbserver-sb\") pod \"dnsmasq-dns-784f69c749-cxssm\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.803013 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-ovsdbserver-sb\") pod \"dnsmasq-dns-784f69c749-cxssm\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.803847 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-ovsdbserver-nb\") pod \"dnsmasq-dns-784f69c749-cxssm\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.803927 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-dns-svc\") pod \"dnsmasq-dns-784f69c749-cxssm\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.804494 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-config\") pod \"dnsmasq-dns-784f69c749-cxssm\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.816247 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-cxssm"] Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.826117 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.857119 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-c9rsj"] Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.858264 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.866338 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.867020 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.868397 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2chcb" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.886804 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-c9rsj"] Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.919282 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-config-data\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.919350 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-scripts\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.919426 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-etc-machine-id\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.919466 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-db-sync-config-data\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.919601 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-combined-ca-bundle\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.919685 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd49x\" (UniqueName: \"kubernetes.io/projected/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-kube-api-access-wd49x\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.925797 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-b6f66b857-h6k98"] Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.927213 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.941698 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2cpx\" (UniqueName: \"kubernetes.io/projected/b29d1a05-edbf-4146-8c0f-21e846bad95f-kube-api-access-p2cpx\") pod \"dnsmasq-dns-784f69c749-cxssm\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.952331 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.952537 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.952806 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.952995 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-gx2rd" Dec 02 01:56:09 crc kubenswrapper[4884]: I1202 01:56:09.957787 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-b6f66b857-h6k98"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.022788 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-scripts\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.023118 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-etc-machine-id\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.023156 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-scripts\") pod \"horizon-b6f66b857-h6k98\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.023185 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-db-sync-config-data\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.023246 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-combined-ca-bundle\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.023268 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp5ss\" (UniqueName: \"kubernetes.io/projected/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-kube-api-access-sp5ss\") pod \"horizon-b6f66b857-h6k98\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.023310 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-config-data\") pod \"horizon-b6f66b857-h6k98\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.023335 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd49x\" (UniqueName: \"kubernetes.io/projected/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-kube-api-access-wd49x\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.023351 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-logs\") pod \"horizon-b6f66b857-h6k98\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.023373 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-horizon-secret-key\") pod \"horizon-b6f66b857-h6k98\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.023405 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-config-data\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.027106 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-etc-machine-id\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.039312 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-db-sync-config-data\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.051057 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-scripts\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.060690 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-combined-ca-bundle\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.067787 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-config-data\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.074991 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.077056 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.081104 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.082661 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.093836 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.125692 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp5ss\" (UniqueName: \"kubernetes.io/projected/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-kube-api-access-sp5ss\") pod \"horizon-b6f66b857-h6k98\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.125765 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-config-data\") pod \"horizon-b6f66b857-h6k98\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.125797 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-logs\") pod \"horizon-b6f66b857-h6k98\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.125814 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-horizon-secret-key\") pod \"horizon-b6f66b857-h6k98\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.125869 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-scripts\") pod \"horizon-b6f66b857-h6k98\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.126536 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-scripts\") pod \"horizon-b6f66b857-h6k98\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.127572 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-config-data\") pod \"horizon-b6f66b857-h6k98\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.127782 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-logs\") pod \"horizon-b6f66b857-h6k98\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.152379 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp5ss\" (UniqueName: \"kubernetes.io/projected/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-kube-api-access-sp5ss\") pod \"horizon-b6f66b857-h6k98\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.152511 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd49x\" (UniqueName: \"kubernetes.io/projected/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-kube-api-access-wd49x\") pod \"cinder-db-sync-c9rsj\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.158639 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-horizon-secret-key\") pod \"horizon-b6f66b857-h6k98\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.168151 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-jn9fg"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.178377 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jn9fg" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.194240 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.195574 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.217100 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8bkvq" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.231270 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmt9v\" (UniqueName: \"kubernetes.io/projected/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-kube-api-access-lmt9v\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.231313 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-scripts\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.231369 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d95342e1-6aff-4feb-9309-26ffa69e786b-config\") pod \"neutron-db-sync-jn9fg\" (UID: \"d95342e1-6aff-4feb-9309-26ffa69e786b\") " pod="openstack/neutron-db-sync-jn9fg" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.250346 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95342e1-6aff-4feb-9309-26ffa69e786b-combined-ca-bundle\") pod \"neutron-db-sync-jn9fg\" (UID: \"d95342e1-6aff-4feb-9309-26ffa69e786b\") " pod="openstack/neutron-db-sync-jn9fg" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.250773 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-run-httpd\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.250806 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.250826 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmk6l\" (UniqueName: \"kubernetes.io/projected/d95342e1-6aff-4feb-9309-26ffa69e786b-kube-api-access-mmk6l\") pod \"neutron-db-sync-jn9fg\" (UID: \"d95342e1-6aff-4feb-9309-26ffa69e786b\") " pod="openstack/neutron-db-sync-jn9fg" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.251967 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-config-data\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.252329 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.252373 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-log-httpd\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.259407 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.259783 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.281209 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jn9fg"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.317725 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.352144 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-5vj4w"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.353898 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.356869 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.357045 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.357301 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-q4l62" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.359960 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d95342e1-6aff-4feb-9309-26ffa69e786b-config\") pod \"neutron-db-sync-jn9fg\" (UID: \"d95342e1-6aff-4feb-9309-26ffa69e786b\") " pod="openstack/neutron-db-sync-jn9fg" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.360051 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95342e1-6aff-4feb-9309-26ffa69e786b-combined-ca-bundle\") pod \"neutron-db-sync-jn9fg\" (UID: \"d95342e1-6aff-4feb-9309-26ffa69e786b\") " pod="openstack/neutron-db-sync-jn9fg" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.360079 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-run-httpd\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.360112 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.360138 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmk6l\" (UniqueName: \"kubernetes.io/projected/d95342e1-6aff-4feb-9309-26ffa69e786b-kube-api-access-mmk6l\") pod \"neutron-db-sync-jn9fg\" (UID: \"d95342e1-6aff-4feb-9309-26ffa69e786b\") " pod="openstack/neutron-db-sync-jn9fg" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.360230 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-config-data\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.360253 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.360288 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-log-httpd\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.360345 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmt9v\" (UniqueName: \"kubernetes.io/projected/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-kube-api-access-lmt9v\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.360372 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-scripts\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.363545 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-run-httpd\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.365581 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-scripts\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.366101 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-log-httpd\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.370482 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-config-data\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.371073 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d95342e1-6aff-4feb-9309-26ffa69e786b-config\") pod \"neutron-db-sync-jn9fg\" (UID: \"d95342e1-6aff-4feb-9309-26ffa69e786b\") " pod="openstack/neutron-db-sync-jn9fg" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.376773 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"0668e3f90b60525b699a422e41510aa58b452367192b181e70798a2fd8381adb"} Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.376813 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"a27f8470a51b318dd1b918b3a885b2506067208c76200336ba07fab79f055264"} Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.380199 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95342e1-6aff-4feb-9309-26ffa69e786b-combined-ca-bundle\") pod \"neutron-db-sync-jn9fg\" (UID: \"d95342e1-6aff-4feb-9309-26ffa69e786b\") " pod="openstack/neutron-db-sync-jn9fg" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.380268 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d75d4f7c7-492fj"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.382908 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.386518 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.390339 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.394919 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmk6l\" (UniqueName: \"kubernetes.io/projected/d95342e1-6aff-4feb-9309-26ffa69e786b-kube-api-access-mmk6l\") pod \"neutron-db-sync-jn9fg\" (UID: \"d95342e1-6aff-4feb-9309-26ffa69e786b\") " pod="openstack/neutron-db-sync-jn9fg" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.409106 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmt9v\" (UniqueName: \"kubernetes.io/projected/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-kube-api-access-lmt9v\") pod \"ceilometer-0\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.417525 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-5vj4w"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.455017 4884 generic.go:334] "Generic (PLEG): container finished" podID="bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21" containerID="cf483adfbc1ccdf1397ec42465607bf68aee7b475be6035b2d90e4bbb3f49416" exitCode=0 Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.455059 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" event={"ID":"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21","Type":"ContainerDied","Data":"cf483adfbc1ccdf1397ec42465607bf68aee7b475be6035b2d90e4bbb3f49416"} Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.502583 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea9acb40-6e2e-4873-9a88-66fb8830d688-logs\") pod \"horizon-7d75d4f7c7-492fj\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.502660 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37412014-287b-42f1-9437-ad164aa34f52-logs\") pod \"placement-db-sync-5vj4w\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.502735 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-config-data\") pod \"placement-db-sync-5vj4w\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.502824 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzjts\" (UniqueName: \"kubernetes.io/projected/ea9acb40-6e2e-4873-9a88-66fb8830d688-kube-api-access-dzjts\") pod \"horizon-7d75d4f7c7-492fj\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.502848 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea9acb40-6e2e-4873-9a88-66fb8830d688-config-data\") pod \"horizon-7d75d4f7c7-492fj\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.502904 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ea9acb40-6e2e-4873-9a88-66fb8830d688-horizon-secret-key\") pod \"horizon-7d75d4f7c7-492fj\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.502970 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-combined-ca-bundle\") pod \"placement-db-sync-5vj4w\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.503083 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbh55\" (UniqueName: \"kubernetes.io/projected/37412014-287b-42f1-9437-ad164aa34f52-kube-api-access-kbh55\") pod \"placement-db-sync-5vj4w\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.503133 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-scripts\") pod \"placement-db-sync-5vj4w\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.503170 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea9acb40-6e2e-4873-9a88-66fb8830d688-scripts\") pod \"horizon-7d75d4f7c7-492fj\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.517805 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-cxssm"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.569450 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-wmhnq"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.571401 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wmhnq" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.575052 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-lmbq5" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.575249 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.581175 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-wmhnq"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.597911 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d75d4f7c7-492fj"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.604916 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbh55\" (UniqueName: \"kubernetes.io/projected/37412014-287b-42f1-9437-ad164aa34f52-kube-api-access-kbh55\") pod \"placement-db-sync-5vj4w\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.604973 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-scripts\") pod \"placement-db-sync-5vj4w\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.605010 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea9acb40-6e2e-4873-9a88-66fb8830d688-scripts\") pod \"horizon-7d75d4f7c7-492fj\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.605067 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea9acb40-6e2e-4873-9a88-66fb8830d688-logs\") pod \"horizon-7d75d4f7c7-492fj\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.605089 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37412014-287b-42f1-9437-ad164aa34f52-logs\") pod \"placement-db-sync-5vj4w\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.605129 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-config-data\") pod \"placement-db-sync-5vj4w\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.605176 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzjts\" (UniqueName: \"kubernetes.io/projected/ea9acb40-6e2e-4873-9a88-66fb8830d688-kube-api-access-dzjts\") pod \"horizon-7d75d4f7c7-492fj\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.605197 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea9acb40-6e2e-4873-9a88-66fb8830d688-config-data\") pod \"horizon-7d75d4f7c7-492fj\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.605235 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ea9acb40-6e2e-4873-9a88-66fb8830d688-horizon-secret-key\") pod \"horizon-7d75d4f7c7-492fj\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.605267 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-combined-ca-bundle\") pod \"placement-db-sync-5vj4w\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.606522 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37412014-287b-42f1-9437-ad164aa34f52-logs\") pod \"placement-db-sync-5vj4w\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.611819 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-gbs55"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.614271 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.620559 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea9acb40-6e2e-4873-9a88-66fb8830d688-config-data\") pod \"horizon-7d75d4f7c7-492fj\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.620700 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea9acb40-6e2e-4873-9a88-66fb8830d688-scripts\") pod \"horizon-7d75d4f7c7-492fj\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.622727 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea9acb40-6e2e-4873-9a88-66fb8830d688-logs\") pod \"horizon-7d75d4f7c7-492fj\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.623350 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-scripts\") pod \"placement-db-sync-5vj4w\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.626528 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-combined-ca-bundle\") pod \"placement-db-sync-5vj4w\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.626649 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-config-data\") pod \"placement-db-sync-5vj4w\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.627123 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ea9acb40-6e2e-4873-9a88-66fb8830d688-horizon-secret-key\") pod \"horizon-7d75d4f7c7-492fj\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.628910 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbh55\" (UniqueName: \"kubernetes.io/projected/37412014-287b-42f1-9437-ad164aa34f52-kube-api-access-kbh55\") pod \"placement-db-sync-5vj4w\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.630959 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzjts\" (UniqueName: \"kubernetes.io/projected/ea9acb40-6e2e-4873-9a88-66fb8830d688-kube-api-access-dzjts\") pod \"horizon-7d75d4f7c7-492fj\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.631885 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-gbs55"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.657870 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.659382 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.663435 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.663563 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.664352 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.664652 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.665031 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.665167 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-rvbnc" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.668062 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.677385 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jn9fg" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.706490 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjdq9\" (UniqueName: \"kubernetes.io/projected/fe8c221e-e320-4358-b940-436f38a423d3-kube-api-access-pjdq9\") pod \"barbican-db-sync-wmhnq\" (UID: \"fe8c221e-e320-4358-b940-436f38a423d3\") " pod="openstack/barbican-db-sync-wmhnq" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.706566 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe8c221e-e320-4358-b940-436f38a423d3-combined-ca-bundle\") pod \"barbican-db-sync-wmhnq\" (UID: \"fe8c221e-e320-4358-b940-436f38a423d3\") " pod="openstack/barbican-db-sync-wmhnq" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.706598 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fe8c221e-e320-4358-b940-436f38a423d3-db-sync-config-data\") pod \"barbican-db-sync-wmhnq\" (UID: \"fe8c221e-e320-4358-b940-436f38a423d3\") " pod="openstack/barbican-db-sync-wmhnq" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.734592 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.772117 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4cqgj"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.790559 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.810052 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-ovsdbserver-nb\") pod \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.810113 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-config\") pod \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.810257 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-ovsdbserver-sb\") pod \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.810277 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-dns-svc\") pod \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.810316 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpwq9\" (UniqueName: \"kubernetes.io/projected/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-kube-api-access-zpwq9\") pod \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\" (UID: \"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21\") " Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.810538 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3058a391-d337-4b74-90c8-9f0d5d81b58e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.810575 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.810591 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.810613 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl976\" (UniqueName: \"kubernetes.io/projected/e47f2b35-d3ff-443a-8f75-37bda0414934-kube-api-access-zl976\") pod \"dnsmasq-dns-f84976bdf-gbs55\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.810649 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-ovsdbserver-sb\") pod \"dnsmasq-dns-f84976bdf-gbs55\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.810695 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjdq9\" (UniqueName: \"kubernetes.io/projected/fe8c221e-e320-4358-b940-436f38a423d3-kube-api-access-pjdq9\") pod \"barbican-db-sync-wmhnq\" (UID: \"fe8c221e-e320-4358-b940-436f38a423d3\") " pod="openstack/barbican-db-sync-wmhnq" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.810797 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlgm9\" (UniqueName: \"kubernetes.io/projected/3058a391-d337-4b74-90c8-9f0d5d81b58e-kube-api-access-jlgm9\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.810829 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-ovsdbserver-nb\") pod \"dnsmasq-dns-f84976bdf-gbs55\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.810849 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-scripts\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.810867 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.812366 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe8c221e-e320-4358-b940-436f38a423d3-combined-ca-bundle\") pod \"barbican-db-sync-wmhnq\" (UID: \"fe8c221e-e320-4358-b940-436f38a423d3\") " pod="openstack/barbican-db-sync-wmhnq" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.812543 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fe8c221e-e320-4358-b940-436f38a423d3-db-sync-config-data\") pod \"barbican-db-sync-wmhnq\" (UID: \"fe8c221e-e320-4358-b940-436f38a423d3\") " pod="openstack/barbican-db-sync-wmhnq" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.812615 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-config-data\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.812653 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-config\") pod \"dnsmasq-dns-f84976bdf-gbs55\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.812688 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-dns-svc\") pod \"dnsmasq-dns-f84976bdf-gbs55\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.812758 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3058a391-d337-4b74-90c8-9f0d5d81b58e-logs\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.816394 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe8c221e-e320-4358-b940-436f38a423d3-combined-ca-bundle\") pod \"barbican-db-sync-wmhnq\" (UID: \"fe8c221e-e320-4358-b940-436f38a423d3\") " pod="openstack/barbican-db-sync-wmhnq" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.816434 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fe8c221e-e320-4358-b940-436f38a423d3-db-sync-config-data\") pod \"barbican-db-sync-wmhnq\" (UID: \"fe8c221e-e320-4358-b940-436f38a423d3\") " pod="openstack/barbican-db-sync-wmhnq" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.825038 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-kube-api-access-zpwq9" (OuterVolumeSpecName: "kube-api-access-zpwq9") pod "bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21" (UID: "bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21"). InnerVolumeSpecName "kube-api-access-zpwq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.838822 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjdq9\" (UniqueName: \"kubernetes.io/projected/fe8c221e-e320-4358-b940-436f38a423d3-kube-api-access-pjdq9\") pod \"barbican-db-sync-wmhnq\" (UID: \"fe8c221e-e320-4358-b940-436f38a423d3\") " pod="openstack/barbican-db-sync-wmhnq" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.862822 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21" (UID: "bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.863485 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21" (UID: "bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.877440 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-config" (OuterVolumeSpecName: "config") pod "bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21" (UID: "bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.895387 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21" (UID: "bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.910759 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wmhnq" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.914605 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-config\") pod \"dnsmasq-dns-f84976bdf-gbs55\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.914651 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-dns-svc\") pod \"dnsmasq-dns-f84976bdf-gbs55\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.914683 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3058a391-d337-4b74-90c8-9f0d5d81b58e-logs\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.914710 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3058a391-d337-4b74-90c8-9f0d5d81b58e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.914752 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.914772 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.914796 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl976\" (UniqueName: \"kubernetes.io/projected/e47f2b35-d3ff-443a-8f75-37bda0414934-kube-api-access-zl976\") pod \"dnsmasq-dns-f84976bdf-gbs55\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.914827 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-ovsdbserver-sb\") pod \"dnsmasq-dns-f84976bdf-gbs55\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.914886 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlgm9\" (UniqueName: \"kubernetes.io/projected/3058a391-d337-4b74-90c8-9f0d5d81b58e-kube-api-access-jlgm9\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.914914 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-ovsdbserver-nb\") pod \"dnsmasq-dns-f84976bdf-gbs55\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.914932 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-scripts\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.914949 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.915002 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-config-data\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.915056 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpwq9\" (UniqueName: \"kubernetes.io/projected/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-kube-api-access-zpwq9\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.915068 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.915079 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.915089 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.915098 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.915335 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3058a391-d337-4b74-90c8-9f0d5d81b58e-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.915719 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-dns-svc\") pod \"dnsmasq-dns-f84976bdf-gbs55\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.915988 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-config\") pod \"dnsmasq-dns-f84976bdf-gbs55\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.916133 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-ovsdbserver-nb\") pod \"dnsmasq-dns-f84976bdf-gbs55\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.916231 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3058a391-d337-4b74-90c8-9f0d5d81b58e-logs\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.916302 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-ovsdbserver-sb\") pod \"dnsmasq-dns-f84976bdf-gbs55\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.918346 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.919075 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.919924 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-scripts\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.922182 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.923795 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-config-data\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.934587 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlgm9\" (UniqueName: \"kubernetes.io/projected/3058a391-d337-4b74-90c8-9f0d5d81b58e-kube-api-access-jlgm9\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.935881 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl976\" (UniqueName: \"kubernetes.io/projected/e47f2b35-d3ff-443a-8f75-37bda0414934-kube-api-access-zl976\") pod \"dnsmasq-dns-f84976bdf-gbs55\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.937149 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-cxssm"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.938794 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.947615 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-c9rsj"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.958470 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.964033 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:56:10 crc kubenswrapper[4884]: E1202 01:56:10.964466 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21" containerName="dnsmasq-dns" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.964477 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21" containerName="dnsmasq-dns" Dec 02 01:56:10 crc kubenswrapper[4884]: E1202 01:56:10.964486 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21" containerName="init" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.964492 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21" containerName="init" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.964663 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21" containerName="dnsmasq-dns" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.965627 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.971491 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.971692 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.973101 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:56:10 crc kubenswrapper[4884]: I1202 01:56:10.991626 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-b6f66b857-h6k98"] Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.023973 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: W1202 01:56:11.074545 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e1fc554_ace1_4c3d_ac7f_5af6b9e932a8.slice/crio-8120ba00419ed548c463db81426bcfced683ba2f086808ecc2e1e0adb944eb70 WatchSource:0}: Error finding container 8120ba00419ed548c463db81426bcfced683ba2f086808ecc2e1e0adb944eb70: Status 404 returned error can't find the container with id 8120ba00419ed548c463db81426bcfced683ba2f086808ecc2e1e0adb944eb70 Dec 02 01:56:11 crc kubenswrapper[4884]: W1202 01:56:11.079876 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0d68bd4_6c59_4e41_b32e_f7918d3b4ada.slice/crio-18df2fffc1924f7ae65d40273225114fb51661eb6360fa097db4957fef32d835 WatchSource:0}: Error finding container 18df2fffc1924f7ae65d40273225114fb51661eb6360fa097db4957fef32d835: Status 404 returned error can't find the container with id 18df2fffc1924f7ae65d40273225114fb51661eb6360fa097db4957fef32d835 Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.117773 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.117867 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05ad28bd-1166-4977-940f-c0d7a041b177-logs\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.117924 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.117951 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tlkv\" (UniqueName: \"kubernetes.io/projected/05ad28bd-1166-4977-940f-c0d7a041b177-kube-api-access-2tlkv\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.117983 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.118015 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-scripts\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.118110 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-config-data\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.118157 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/05ad28bd-1166-4977-940f-c0d7a041b177-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.219961 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/05ad28bd-1166-4977-940f-c0d7a041b177-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.220232 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.220306 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05ad28bd-1166-4977-940f-c0d7a041b177-logs\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.220342 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.220358 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tlkv\" (UniqueName: \"kubernetes.io/projected/05ad28bd-1166-4977-940f-c0d7a041b177-kube-api-access-2tlkv\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.220377 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.220402 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-scripts\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.220436 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-config-data\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.220452 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/05ad28bd-1166-4977-940f-c0d7a041b177-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.221241 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.225634 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-config-data\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.227381 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.227638 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05ad28bd-1166-4977-940f-c0d7a041b177-logs\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.232497 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.251129 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-scripts\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.254529 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tlkv\" (UniqueName: \"kubernetes.io/projected/05ad28bd-1166-4977-940f-c0d7a041b177-kube-api-access-2tlkv\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.286192 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.469280 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784f69c749-cxssm" event={"ID":"b29d1a05-edbf-4146-8c0f-21e846bad95f","Type":"ContainerStarted","Data":"2564426c85397a6033e64c4492811ca9bd931f7bb712b4687bb1328390c70ee3"} Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.476465 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" event={"ID":"bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21","Type":"ContainerDied","Data":"efb1a6f5eb68f020ddcdf4797d49066fbe59f9c0819e72fc2c13549717bc49d1"} Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.476518 4884 scope.go:117] "RemoveContainer" containerID="cf483adfbc1ccdf1397ec42465607bf68aee7b475be6035b2d90e4bbb3f49416" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.476606 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-mltjh" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.493938 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4cqgj" event={"ID":"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8","Type":"ContainerStarted","Data":"8120ba00419ed548c463db81426bcfced683ba2f086808ecc2e1e0adb944eb70"} Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.496905 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b6f66b857-h6k98" event={"ID":"d0550743-d9cf-4b93-a2df-86bd40d2dd3d","Type":"ContainerStarted","Data":"6c87bb72dcade19b2a9a857c8cc38b0e51b4b38933fc5f44e251dd9b713657a6"} Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.499380 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-c9rsj" event={"ID":"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada","Type":"ContainerStarted","Data":"18df2fffc1924f7ae65d40273225114fb51661eb6360fa097db4957fef32d835"} Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.507122 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"50d70edca288db03f1313ccea44e5cbd6ccc39e9ac7d4e9d11e58ad1ba0c8797"} Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.522464 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-mltjh"] Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.525071 4884 scope.go:117] "RemoveContainer" containerID="b216246e9d169813390faa7f6252303d40feb3b5bb9be26c0fbd7371d103fb20" Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.539797 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-mltjh"] Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.574121 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:56:11 crc kubenswrapper[4884]: I1202 01:56:11.590310 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:11.681778 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21" path="/var/lib/kubelet/pods/bc347dd5-89f1-4f68-8d3b-c8fedaaa3c21/volumes" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:11.770102 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-5vj4w"] Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:11.777149 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jn9fg"] Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.256477 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.307936 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-b6f66b857-h6k98"] Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.337433 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-64d85fdcb9-ncfgl"] Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.352699 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.365532 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.378679 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64d85fdcb9-ncfgl"] Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.452777 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f9d2899-4e00-467e-afbe-214f40042543-horizon-secret-key\") pod \"horizon-64d85fdcb9-ncfgl\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.453656 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f9d2899-4e00-467e-afbe-214f40042543-logs\") pod \"horizon-64d85fdcb9-ncfgl\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.453805 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd55b\" (UniqueName: \"kubernetes.io/projected/8f9d2899-4e00-467e-afbe-214f40042543-kube-api-access-jd55b\") pod \"horizon-64d85fdcb9-ncfgl\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.453852 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f9d2899-4e00-467e-afbe-214f40042543-config-data\") pod \"horizon-64d85fdcb9-ncfgl\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.453899 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f9d2899-4e00-467e-afbe-214f40042543-scripts\") pod \"horizon-64d85fdcb9-ncfgl\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.528725 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jn9fg" event={"ID":"d95342e1-6aff-4feb-9309-26ffa69e786b","Type":"ContainerStarted","Data":"a5d68bf6080508333b0425d2b369cbf613c67a868be4a5af9d924e71ccbbd979"} Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.528773 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jn9fg" event={"ID":"d95342e1-6aff-4feb-9309-26ffa69e786b","Type":"ContainerStarted","Data":"4f1e5c894cb8298a0eae5ae62b12d4575ba875e1b943cac586ece27c6ea21f73"} Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.546524 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-jn9fg" podStartSLOduration=3.546507459 podStartE2EDuration="3.546507459s" podCreationTimestamp="2025-12-02 01:56:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:12.542564791 +0000 UTC m=+1189.218401685" watchObservedRunningTime="2025-12-02 01:56:12.546507459 +0000 UTC m=+1189.222344343" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.560592 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f9d2899-4e00-467e-afbe-214f40042543-logs\") pod \"horizon-64d85fdcb9-ncfgl\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.560651 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd55b\" (UniqueName: \"kubernetes.io/projected/8f9d2899-4e00-467e-afbe-214f40042543-kube-api-access-jd55b\") pod \"horizon-64d85fdcb9-ncfgl\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.560673 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f9d2899-4e00-467e-afbe-214f40042543-config-data\") pod \"horizon-64d85fdcb9-ncfgl\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.560700 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f9d2899-4e00-467e-afbe-214f40042543-scripts\") pod \"horizon-64d85fdcb9-ncfgl\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.560726 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f9d2899-4e00-467e-afbe-214f40042543-horizon-secret-key\") pod \"horizon-64d85fdcb9-ncfgl\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.561694 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"8dc2a1f5-a7a6-4ce5-a711-77455efe2224","Type":"ContainerStarted","Data":"c0f86044531bf5c357bd8e588f80968e6c2c71a605c9c698d7204b09c7dd9000"} Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.564067 4884 generic.go:334] "Generic (PLEG): container finished" podID="b29d1a05-edbf-4146-8c0f-21e846bad95f" containerID="4fc6d3dda795e664098198f1d7691a326528d7635c48af7b42e4bf234f55d1af" exitCode=0 Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.564110 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784f69c749-cxssm" event={"ID":"b29d1a05-edbf-4146-8c0f-21e846bad95f","Type":"ContainerDied","Data":"4fc6d3dda795e664098198f1d7691a326528d7635c48af7b42e4bf234f55d1af"} Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.567984 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f9d2899-4e00-467e-afbe-214f40042543-horizon-secret-key\") pod \"horizon-64d85fdcb9-ncfgl\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.571717 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f9d2899-4e00-467e-afbe-214f40042543-scripts\") pod \"horizon-64d85fdcb9-ncfgl\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.571933 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f9d2899-4e00-467e-afbe-214f40042543-logs\") pod \"horizon-64d85fdcb9-ncfgl\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.572085 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f9d2899-4e00-467e-afbe-214f40042543-config-data\") pod \"horizon-64d85fdcb9-ncfgl\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.576337 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5vj4w" event={"ID":"37412014-287b-42f1-9437-ad164aa34f52","Type":"ContainerStarted","Data":"b3f53e3e6bc800ed5e6e0b38aa6149910b413086b3aab804c20fc459fc543c04"} Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.585277 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4cqgj" event={"ID":"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8","Type":"ContainerStarted","Data":"c3a7e3020f57948047ca21a9940c2734138a0269a3009feb28ee61e353d08eb8"} Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.595233 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45","Type":"ContainerStarted","Data":"6d7bdda8240d78eef550bf103054c243a7691255acc6b3f4211e1269eaddd053"} Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.599984 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd55b\" (UniqueName: \"kubernetes.io/projected/8f9d2899-4e00-467e-afbe-214f40042543-kube-api-access-jd55b\") pod \"horizon-64d85fdcb9-ncfgl\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.644607 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=40.06678203 podStartE2EDuration="1m1.64458631s" podCreationTimestamp="2025-12-02 01:55:11 +0000 UTC" firstStartedPulling="2025-12-02 01:55:46.509371792 +0000 UTC m=+1163.185208676" lastFinishedPulling="2025-12-02 01:56:08.087176062 +0000 UTC m=+1184.763012956" observedRunningTime="2025-12-02 01:56:12.621835528 +0000 UTC m=+1189.297672422" watchObservedRunningTime="2025-12-02 01:56:12.64458631 +0000 UTC m=+1189.320423194" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.670784 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-4cqgj" podStartSLOduration=3.670734756 podStartE2EDuration="3.670734756s" podCreationTimestamp="2025-12-02 01:56:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:12.654494315 +0000 UTC m=+1189.330331189" watchObservedRunningTime="2025-12-02 01:56:12.670734756 +0000 UTC m=+1189.346571640" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.681998 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.724263 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-wmhnq"] Dec 02 01:56:12 crc kubenswrapper[4884]: W1202 01:56:12.733735 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode47f2b35_d3ff_443a_8f75_37bda0414934.slice/crio-4e40349ff4d57695dd26333b2ce2a8c4e29b720662d8a020e715824a266b1b11 WatchSource:0}: Error finding container 4e40349ff4d57695dd26333b2ce2a8c4e29b720662d8a020e715824a266b1b11: Status 404 returned error can't find the container with id 4e40349ff4d57695dd26333b2ce2a8c4e29b720662d8a020e715824a266b1b11 Dec 02 01:56:12 crc kubenswrapper[4884]: W1202 01:56:12.734167 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea9acb40_6e2e_4873_9a88_66fb8830d688.slice/crio-043624d531af668431812d880ea3320f4866dd8182efabc11e517cbe5c9452be WatchSource:0}: Error finding container 043624d531af668431812d880ea3320f4866dd8182efabc11e517cbe5c9452be: Status 404 returned error can't find the container with id 043624d531af668431812d880ea3320f4866dd8182efabc11e517cbe5c9452be Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.740662 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-gbs55"] Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.763726 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d75d4f7c7-492fj"] Dec 02 01:56:12 crc kubenswrapper[4884]: W1202 01:56:12.776301 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe8c221e_e320_4358_b940_436f38a423d3.slice/crio-7c3d69fd23b6cc1507a9b34b533179bac25cd94f108362939294514e1ab412f1 WatchSource:0}: Error finding container 7c3d69fd23b6cc1507a9b34b533179bac25cd94f108362939294514e1ab412f1: Status 404 returned error can't find the container with id 7c3d69fd23b6cc1507a9b34b533179bac25cd94f108362939294514e1ab412f1 Dec 02 01:56:12 crc kubenswrapper[4884]: I1202 01:56:12.803388 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.009597 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.019206 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-gbs55"] Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.058472 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-crspk"] Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.059985 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.063006 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.074167 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-crspk"] Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.191760 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.191802 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.191836 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn5hp\" (UniqueName: \"kubernetes.io/projected/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-kube-api-access-fn5hp\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.191906 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.191947 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.191967 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-config\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.194625 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.295182 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.295274 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-config\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.295299 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.295378 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.295409 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.295446 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn5hp\" (UniqueName: \"kubernetes.io/projected/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-kube-api-access-fn5hp\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.296161 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.296219 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-config\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.296306 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.296583 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.296724 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.333773 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn5hp\" (UniqueName: \"kubernetes.io/projected/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-kube-api-access-fn5hp\") pod \"dnsmasq-dns-785d8bcb8c-crspk\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.348049 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64d85fdcb9-ncfgl"] Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.393813 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.400844 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.499079 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-ovsdbserver-nb\") pod \"b29d1a05-edbf-4146-8c0f-21e846bad95f\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.499217 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-ovsdbserver-sb\") pod \"b29d1a05-edbf-4146-8c0f-21e846bad95f\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.499275 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2cpx\" (UniqueName: \"kubernetes.io/projected/b29d1a05-edbf-4146-8c0f-21e846bad95f-kube-api-access-p2cpx\") pod \"b29d1a05-edbf-4146-8c0f-21e846bad95f\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.499353 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-dns-svc\") pod \"b29d1a05-edbf-4146-8c0f-21e846bad95f\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.499421 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-config\") pod \"b29d1a05-edbf-4146-8c0f-21e846bad95f\" (UID: \"b29d1a05-edbf-4146-8c0f-21e846bad95f\") " Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.509481 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b29d1a05-edbf-4146-8c0f-21e846bad95f-kube-api-access-p2cpx" (OuterVolumeSpecName: "kube-api-access-p2cpx") pod "b29d1a05-edbf-4146-8c0f-21e846bad95f" (UID: "b29d1a05-edbf-4146-8c0f-21e846bad95f"). InnerVolumeSpecName "kube-api-access-p2cpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.530588 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-config" (OuterVolumeSpecName: "config") pod "b29d1a05-edbf-4146-8c0f-21e846bad95f" (UID: "b29d1a05-edbf-4146-8c0f-21e846bad95f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.536375 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b29d1a05-edbf-4146-8c0f-21e846bad95f" (UID: "b29d1a05-edbf-4146-8c0f-21e846bad95f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.555141 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b29d1a05-edbf-4146-8c0f-21e846bad95f" (UID: "b29d1a05-edbf-4146-8c0f-21e846bad95f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.555896 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b29d1a05-edbf-4146-8c0f-21e846bad95f" (UID: "b29d1a05-edbf-4146-8c0f-21e846bad95f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.602366 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.602398 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2cpx\" (UniqueName: \"kubernetes.io/projected/b29d1a05-edbf-4146-8c0f-21e846bad95f-kube-api-access-p2cpx\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.602409 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.602418 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.602427 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b29d1a05-edbf-4146-8c0f-21e846bad95f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.610266 4884 generic.go:334] "Generic (PLEG): container finished" podID="e47f2b35-d3ff-443a-8f75-37bda0414934" containerID="999da56f259425355183d9689298a39b9230fa7cc0b03e2753298e9f7028369b" exitCode=0 Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.610333 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-gbs55" event={"ID":"e47f2b35-d3ff-443a-8f75-37bda0414934","Type":"ContainerDied","Data":"999da56f259425355183d9689298a39b9230fa7cc0b03e2753298e9f7028369b"} Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.610408 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-gbs55" event={"ID":"e47f2b35-d3ff-443a-8f75-37bda0414934","Type":"ContainerStarted","Data":"4e40349ff4d57695dd26333b2ce2a8c4e29b720662d8a020e715824a266b1b11"} Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.616294 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3058a391-d337-4b74-90c8-9f0d5d81b58e","Type":"ContainerStarted","Data":"78d670f49fe5008b784559b3a5e75ac857b63085c5995dec6db1a816be6b47a1"} Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.617484 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wmhnq" event={"ID":"fe8c221e-e320-4358-b940-436f38a423d3","Type":"ContainerStarted","Data":"7c3d69fd23b6cc1507a9b34b533179bac25cd94f108362939294514e1ab412f1"} Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.646364 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"05ad28bd-1166-4977-940f-c0d7a041b177","Type":"ContainerStarted","Data":"72925fbb72246bd1e99e11c3228f3f8cc34b19ca47130c3d1662454ce251b016"} Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.676757 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64d85fdcb9-ncfgl" event={"ID":"8f9d2899-4e00-467e-afbe-214f40042543","Type":"ContainerStarted","Data":"e3e9b26d58145e3c24fa719a8445d86e9ff72fd967c9ee1adb0c73b71391440e"} Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.682858 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d75d4f7c7-492fj" event={"ID":"ea9acb40-6e2e-4873-9a88-66fb8830d688","Type":"ContainerStarted","Data":"043624d531af668431812d880ea3320f4866dd8182efabc11e517cbe5c9452be"} Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.687819 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-784f69c749-cxssm" event={"ID":"b29d1a05-edbf-4146-8c0f-21e846bad95f","Type":"ContainerDied","Data":"2564426c85397a6033e64c4492811ca9bd931f7bb712b4687bb1328390c70ee3"} Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.687893 4884 scope.go:117] "RemoveContainer" containerID="4fc6d3dda795e664098198f1d7691a326528d7635c48af7b42e4bf234f55d1af" Dec 02 01:56:13 crc kubenswrapper[4884]: I1202 01:56:13.688143 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-784f69c749-cxssm" Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.093574 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-cxssm"] Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.143655 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-784f69c749-cxssm"] Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.187580 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-crspk"] Dec 02 01:56:14 crc kubenswrapper[4884]: W1202 01:56:14.228774 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f6fe6c3_c15e_45d3_91e5_1474030f2bbd.slice/crio-038dd18c1e37350d8c154f41719ca2df601d7ef0bcd8c1e2f405f1491caa4127 WatchSource:0}: Error finding container 038dd18c1e37350d8c154f41719ca2df601d7ef0bcd8c1e2f405f1491caa4127: Status 404 returned error can't find the container with id 038dd18c1e37350d8c154f41719ca2df601d7ef0bcd8c1e2f405f1491caa4127 Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.406330 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.537141 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-config\") pod \"e47f2b35-d3ff-443a-8f75-37bda0414934\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.537572 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-ovsdbserver-nb\") pod \"e47f2b35-d3ff-443a-8f75-37bda0414934\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.537612 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-dns-svc\") pod \"e47f2b35-d3ff-443a-8f75-37bda0414934\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.537657 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl976\" (UniqueName: \"kubernetes.io/projected/e47f2b35-d3ff-443a-8f75-37bda0414934-kube-api-access-zl976\") pod \"e47f2b35-d3ff-443a-8f75-37bda0414934\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.537789 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-ovsdbserver-sb\") pod \"e47f2b35-d3ff-443a-8f75-37bda0414934\" (UID: \"e47f2b35-d3ff-443a-8f75-37bda0414934\") " Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.551642 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e47f2b35-d3ff-443a-8f75-37bda0414934-kube-api-access-zl976" (OuterVolumeSpecName: "kube-api-access-zl976") pod "e47f2b35-d3ff-443a-8f75-37bda0414934" (UID: "e47f2b35-d3ff-443a-8f75-37bda0414934"). InnerVolumeSpecName "kube-api-access-zl976". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.559174 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-config" (OuterVolumeSpecName: "config") pod "e47f2b35-d3ff-443a-8f75-37bda0414934" (UID: "e47f2b35-d3ff-443a-8f75-37bda0414934"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.564164 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e47f2b35-d3ff-443a-8f75-37bda0414934" (UID: "e47f2b35-d3ff-443a-8f75-37bda0414934"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.584448 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e47f2b35-d3ff-443a-8f75-37bda0414934" (UID: "e47f2b35-d3ff-443a-8f75-37bda0414934"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.603879 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e47f2b35-d3ff-443a-8f75-37bda0414934" (UID: "e47f2b35-d3ff-443a-8f75-37bda0414934"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.639967 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.667414 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.667427 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.667437 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e47f2b35-d3ff-443a-8f75-37bda0414934-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.667447 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl976\" (UniqueName: \"kubernetes.io/projected/e47f2b35-d3ff-443a-8f75-37bda0414934-kube-api-access-zl976\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.710078 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84976bdf-gbs55" event={"ID":"e47f2b35-d3ff-443a-8f75-37bda0414934","Type":"ContainerDied","Data":"4e40349ff4d57695dd26333b2ce2a8c4e29b720662d8a020e715824a266b1b11"} Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.710137 4884 scope.go:117] "RemoveContainer" containerID="999da56f259425355183d9689298a39b9230fa7cc0b03e2753298e9f7028369b" Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.710101 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84976bdf-gbs55" Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.716862 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3058a391-d337-4b74-90c8-9f0d5d81b58e","Type":"ContainerStarted","Data":"7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239"} Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.720370 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"05ad28bd-1166-4977-940f-c0d7a041b177","Type":"ContainerStarted","Data":"e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce"} Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.728450 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" event={"ID":"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd","Type":"ContainerStarted","Data":"544956c368c052542d947e912d089b6be064eddfad9ce9e6c447e3200a3cf0e8"} Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.728483 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" event={"ID":"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd","Type":"ContainerStarted","Data":"038dd18c1e37350d8c154f41719ca2df601d7ef0bcd8c1e2f405f1491caa4127"} Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.816238 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-gbs55"] Dec 02 01:56:14 crc kubenswrapper[4884]: I1202 01:56:14.841322 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f84976bdf-gbs55"] Dec 02 01:56:15 crc kubenswrapper[4884]: I1202 01:56:15.631042 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b29d1a05-edbf-4146-8c0f-21e846bad95f" path="/var/lib/kubelet/pods/b29d1a05-edbf-4146-8c0f-21e846bad95f/volumes" Dec 02 01:56:15 crc kubenswrapper[4884]: I1202 01:56:15.631538 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e47f2b35-d3ff-443a-8f75-37bda0414934" path="/var/lib/kubelet/pods/e47f2b35-d3ff-443a-8f75-37bda0414934/volumes" Dec 02 01:56:15 crc kubenswrapper[4884]: I1202 01:56:15.748534 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3058a391-d337-4b74-90c8-9f0d5d81b58e","Type":"ContainerStarted","Data":"33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8"} Dec 02 01:56:15 crc kubenswrapper[4884]: I1202 01:56:15.748668 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3058a391-d337-4b74-90c8-9f0d5d81b58e" containerName="glance-log" containerID="cri-o://7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239" gracePeriod=30 Dec 02 01:56:15 crc kubenswrapper[4884]: I1202 01:56:15.749111 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3058a391-d337-4b74-90c8-9f0d5d81b58e" containerName="glance-httpd" containerID="cri-o://33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8" gracePeriod=30 Dec 02 01:56:15 crc kubenswrapper[4884]: I1202 01:56:15.752030 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"05ad28bd-1166-4977-940f-c0d7a041b177","Type":"ContainerStarted","Data":"ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef"} Dec 02 01:56:15 crc kubenswrapper[4884]: I1202 01:56:15.752107 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="05ad28bd-1166-4977-940f-c0d7a041b177" containerName="glance-log" containerID="cri-o://e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce" gracePeriod=30 Dec 02 01:56:15 crc kubenswrapper[4884]: I1202 01:56:15.752141 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="05ad28bd-1166-4977-940f-c0d7a041b177" containerName="glance-httpd" containerID="cri-o://ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef" gracePeriod=30 Dec 02 01:56:15 crc kubenswrapper[4884]: I1202 01:56:15.754800 4884 generic.go:334] "Generic (PLEG): container finished" podID="4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" containerID="544956c368c052542d947e912d089b6be064eddfad9ce9e6c447e3200a3cf0e8" exitCode=0 Dec 02 01:56:15 crc kubenswrapper[4884]: I1202 01:56:15.754853 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" event={"ID":"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd","Type":"ContainerDied","Data":"544956c368c052542d947e912d089b6be064eddfad9ce9e6c447e3200a3cf0e8"} Dec 02 01:56:15 crc kubenswrapper[4884]: I1202 01:56:15.765906 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.765889693 podStartE2EDuration="5.765889693s" podCreationTimestamp="2025-12-02 01:56:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:15.763408551 +0000 UTC m=+1192.439245445" watchObservedRunningTime="2025-12-02 01:56:15.765889693 +0000 UTC m=+1192.441726567" Dec 02 01:56:15 crc kubenswrapper[4884]: I1202 01:56:15.778608 4884 generic.go:334] "Generic (PLEG): container finished" podID="5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8" containerID="c3a7e3020f57948047ca21a9940c2734138a0269a3009feb28ee61e353d08eb8" exitCode=0 Dec 02 01:56:15 crc kubenswrapper[4884]: I1202 01:56:15.778647 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4cqgj" event={"ID":"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8","Type":"ContainerDied","Data":"c3a7e3020f57948047ca21a9940c2734138a0269a3009feb28ee61e353d08eb8"} Dec 02 01:56:15 crc kubenswrapper[4884]: I1202 01:56:15.844481 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.844458292 podStartE2EDuration="6.844458292s" podCreationTimestamp="2025-12-02 01:56:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:15.811414477 +0000 UTC m=+1192.487251361" watchObservedRunningTime="2025-12-02 01:56:15.844458292 +0000 UTC m=+1192.520295176" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.431200 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.532986 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.602504 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlgm9\" (UniqueName: \"kubernetes.io/projected/3058a391-d337-4b74-90c8-9f0d5d81b58e-kube-api-access-jlgm9\") pod \"3058a391-d337-4b74-90c8-9f0d5d81b58e\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.602641 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-public-tls-certs\") pod \"3058a391-d337-4b74-90c8-9f0d5d81b58e\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.602692 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-combined-ca-bundle\") pod \"3058a391-d337-4b74-90c8-9f0d5d81b58e\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.602821 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-config-data\") pod \"3058a391-d337-4b74-90c8-9f0d5d81b58e\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.602935 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3058a391-d337-4b74-90c8-9f0d5d81b58e-logs\") pod \"3058a391-d337-4b74-90c8-9f0d5d81b58e\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.602983 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3058a391-d337-4b74-90c8-9f0d5d81b58e-httpd-run\") pod \"3058a391-d337-4b74-90c8-9f0d5d81b58e\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.603015 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-scripts\") pod \"3058a391-d337-4b74-90c8-9f0d5d81b58e\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.603240 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"3058a391-d337-4b74-90c8-9f0d5d81b58e\" (UID: \"3058a391-d337-4b74-90c8-9f0d5d81b58e\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.603807 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3058a391-d337-4b74-90c8-9f0d5d81b58e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3058a391-d337-4b74-90c8-9f0d5d81b58e" (UID: "3058a391-d337-4b74-90c8-9f0d5d81b58e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.603991 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3058a391-d337-4b74-90c8-9f0d5d81b58e-logs" (OuterVolumeSpecName: "logs") pod "3058a391-d337-4b74-90c8-9f0d5d81b58e" (UID: "3058a391-d337-4b74-90c8-9f0d5d81b58e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.609031 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-scripts" (OuterVolumeSpecName: "scripts") pod "3058a391-d337-4b74-90c8-9f0d5d81b58e" (UID: "3058a391-d337-4b74-90c8-9f0d5d81b58e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.609091 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3058a391-d337-4b74-90c8-9f0d5d81b58e-kube-api-access-jlgm9" (OuterVolumeSpecName: "kube-api-access-jlgm9") pod "3058a391-d337-4b74-90c8-9f0d5d81b58e" (UID: "3058a391-d337-4b74-90c8-9f0d5d81b58e"). InnerVolumeSpecName "kube-api-access-jlgm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.609160 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "3058a391-d337-4b74-90c8-9f0d5d81b58e" (UID: "3058a391-d337-4b74-90c8-9f0d5d81b58e"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.638433 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3058a391-d337-4b74-90c8-9f0d5d81b58e" (UID: "3058a391-d337-4b74-90c8-9f0d5d81b58e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.665046 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3058a391-d337-4b74-90c8-9f0d5d81b58e" (UID: "3058a391-d337-4b74-90c8-9f0d5d81b58e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.665060 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-config-data" (OuterVolumeSpecName: "config-data") pod "3058a391-d337-4b74-90c8-9f0d5d81b58e" (UID: "3058a391-d337-4b74-90c8-9f0d5d81b58e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.704520 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05ad28bd-1166-4977-940f-c0d7a041b177-logs\") pod \"05ad28bd-1166-4977-940f-c0d7a041b177\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.704562 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"05ad28bd-1166-4977-940f-c0d7a041b177\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.704602 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-internal-tls-certs\") pod \"05ad28bd-1166-4977-940f-c0d7a041b177\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.704678 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/05ad28bd-1166-4977-940f-c0d7a041b177-httpd-run\") pod \"05ad28bd-1166-4977-940f-c0d7a041b177\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.704825 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tlkv\" (UniqueName: \"kubernetes.io/projected/05ad28bd-1166-4977-940f-c0d7a041b177-kube-api-access-2tlkv\") pod \"05ad28bd-1166-4977-940f-c0d7a041b177\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.704922 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-scripts\") pod \"05ad28bd-1166-4977-940f-c0d7a041b177\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.704940 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-config-data\") pod \"05ad28bd-1166-4977-940f-c0d7a041b177\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.704980 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-combined-ca-bundle\") pod \"05ad28bd-1166-4977-940f-c0d7a041b177\" (UID: \"05ad28bd-1166-4977-940f-c0d7a041b177\") " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.705181 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05ad28bd-1166-4977-940f-c0d7a041b177-logs" (OuterVolumeSpecName: "logs") pod "05ad28bd-1166-4977-940f-c0d7a041b177" (UID: "05ad28bd-1166-4977-940f-c0d7a041b177"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.705426 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.705440 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3058a391-d337-4b74-90c8-9f0d5d81b58e-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.705449 4884 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3058a391-d337-4b74-90c8-9f0d5d81b58e-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.705457 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.705477 4884 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.705486 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlgm9\" (UniqueName: \"kubernetes.io/projected/3058a391-d337-4b74-90c8-9f0d5d81b58e-kube-api-access-jlgm9\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.705495 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05ad28bd-1166-4977-940f-c0d7a041b177-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.705503 4884 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.705512 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3058a391-d337-4b74-90c8-9f0d5d81b58e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.705722 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05ad28bd-1166-4977-940f-c0d7a041b177-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "05ad28bd-1166-4977-940f-c0d7a041b177" (UID: "05ad28bd-1166-4977-940f-c0d7a041b177"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.709982 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "05ad28bd-1166-4977-940f-c0d7a041b177" (UID: "05ad28bd-1166-4977-940f-c0d7a041b177"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.711989 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05ad28bd-1166-4977-940f-c0d7a041b177-kube-api-access-2tlkv" (OuterVolumeSpecName: "kube-api-access-2tlkv") pod "05ad28bd-1166-4977-940f-c0d7a041b177" (UID: "05ad28bd-1166-4977-940f-c0d7a041b177"). InnerVolumeSpecName "kube-api-access-2tlkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.712887 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-scripts" (OuterVolumeSpecName: "scripts") pod "05ad28bd-1166-4977-940f-c0d7a041b177" (UID: "05ad28bd-1166-4977-940f-c0d7a041b177"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.728559 4884 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.734081 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05ad28bd-1166-4977-940f-c0d7a041b177" (UID: "05ad28bd-1166-4977-940f-c0d7a041b177"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.760834 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "05ad28bd-1166-4977-940f-c0d7a041b177" (UID: "05ad28bd-1166-4977-940f-c0d7a041b177"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.772835 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-config-data" (OuterVolumeSpecName: "config-data") pod "05ad28bd-1166-4977-940f-c0d7a041b177" (UID: "05ad28bd-1166-4977-940f-c0d7a041b177"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.795267 4884 generic.go:334] "Generic (PLEG): container finished" podID="3058a391-d337-4b74-90c8-9f0d5d81b58e" containerID="33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8" exitCode=0 Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.795294 4884 generic.go:334] "Generic (PLEG): container finished" podID="3058a391-d337-4b74-90c8-9f0d5d81b58e" containerID="7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239" exitCode=143 Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.795323 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.795344 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3058a391-d337-4b74-90c8-9f0d5d81b58e","Type":"ContainerDied","Data":"33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8"} Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.795394 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3058a391-d337-4b74-90c8-9f0d5d81b58e","Type":"ContainerDied","Data":"7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239"} Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.795406 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3058a391-d337-4b74-90c8-9f0d5d81b58e","Type":"ContainerDied","Data":"78d670f49fe5008b784559b3a5e75ac857b63085c5995dec6db1a816be6b47a1"} Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.795431 4884 scope.go:117] "RemoveContainer" containerID="33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.798857 4884 generic.go:334] "Generic (PLEG): container finished" podID="05ad28bd-1166-4977-940f-c0d7a041b177" containerID="ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef" exitCode=0 Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.799414 4884 generic.go:334] "Generic (PLEG): container finished" podID="05ad28bd-1166-4977-940f-c0d7a041b177" containerID="e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce" exitCode=143 Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.798926 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.798954 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"05ad28bd-1166-4977-940f-c0d7a041b177","Type":"ContainerDied","Data":"ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef"} Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.799527 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"05ad28bd-1166-4977-940f-c0d7a041b177","Type":"ContainerDied","Data":"e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce"} Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.799546 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"05ad28bd-1166-4977-940f-c0d7a041b177","Type":"ContainerDied","Data":"72925fbb72246bd1e99e11c3228f3f8cc34b19ca47130c3d1662454ce251b016"} Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.802881 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" event={"ID":"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd","Type":"ContainerStarted","Data":"8158cc45cc0365e890619d1e6f8e9ada649c3e8ae14caa6a07bd7153ed7ecd78"} Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.802941 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.806844 4884 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.806878 4884 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/05ad28bd-1166-4977-940f-c0d7a041b177-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.806892 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tlkv\" (UniqueName: \"kubernetes.io/projected/05ad28bd-1166-4977-940f-c0d7a041b177-kube-api-access-2tlkv\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.806907 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.806919 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.806930 4884 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.806942 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05ad28bd-1166-4977-940f-c0d7a041b177-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.807003 4884 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.827874 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" podStartSLOduration=3.8278593709999997 podStartE2EDuration="3.827859371s" podCreationTimestamp="2025-12-02 01:56:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:16.81850702 +0000 UTC m=+1193.494343904" watchObservedRunningTime="2025-12-02 01:56:16.827859371 +0000 UTC m=+1193.503696255" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.836439 4884 scope.go:117] "RemoveContainer" containerID="7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.873137 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.881808 4884 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.943104 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.957347 4884 scope.go:117] "RemoveContainer" containerID="33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.958042 4884 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:16 crc kubenswrapper[4884]: E1202 01:56:16.961069 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8\": container with ID starting with 33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8 not found: ID does not exist" containerID="33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.961106 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8"} err="failed to get container status \"33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8\": rpc error: code = NotFound desc = could not find container \"33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8\": container with ID starting with 33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8 not found: ID does not exist" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.961134 4884 scope.go:117] "RemoveContainer" containerID="7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239" Dec 02 01:56:16 crc kubenswrapper[4884]: E1202 01:56:16.962353 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239\": container with ID starting with 7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239 not found: ID does not exist" containerID="7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.962388 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239"} err="failed to get container status \"7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239\": rpc error: code = NotFound desc = could not find container \"7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239\": container with ID starting with 7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239 not found: ID does not exist" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.962415 4884 scope.go:117] "RemoveContainer" containerID="33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.964856 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8"} err="failed to get container status \"33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8\": rpc error: code = NotFound desc = could not find container \"33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8\": container with ID starting with 33b340fe8ff1c5b8755f9095c95488c1091a86be2ad1396b60cfca1aa6d477b8 not found: ID does not exist" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.964891 4884 scope.go:117] "RemoveContainer" containerID="7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.965461 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239"} err="failed to get container status \"7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239\": rpc error: code = NotFound desc = could not find container \"7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239\": container with ID starting with 7279e75786201584f3cf06f4bfb52d9f1482afe1ce26438cd444c16b4a533239 not found: ID does not exist" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.965481 4884 scope.go:117] "RemoveContainer" containerID="ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.966888 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.971850 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.971893 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.978297 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.987645 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:56:16 crc kubenswrapper[4884]: E1202 01:56:16.988073 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05ad28bd-1166-4977-940f-c0d7a041b177" containerName="glance-httpd" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.988091 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="05ad28bd-1166-4977-940f-c0d7a041b177" containerName="glance-httpd" Dec 02 01:56:16 crc kubenswrapper[4884]: E1202 01:56:16.988117 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b29d1a05-edbf-4146-8c0f-21e846bad95f" containerName="init" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.988123 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="b29d1a05-edbf-4146-8c0f-21e846bad95f" containerName="init" Dec 02 01:56:16 crc kubenswrapper[4884]: E1202 01:56:16.988137 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3058a391-d337-4b74-90c8-9f0d5d81b58e" containerName="glance-log" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.988143 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="3058a391-d337-4b74-90c8-9f0d5d81b58e" containerName="glance-log" Dec 02 01:56:16 crc kubenswrapper[4884]: E1202 01:56:16.988154 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3058a391-d337-4b74-90c8-9f0d5d81b58e" containerName="glance-httpd" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.988162 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="3058a391-d337-4b74-90c8-9f0d5d81b58e" containerName="glance-httpd" Dec 02 01:56:16 crc kubenswrapper[4884]: E1202 01:56:16.988176 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e47f2b35-d3ff-443a-8f75-37bda0414934" containerName="init" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.988184 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="e47f2b35-d3ff-443a-8f75-37bda0414934" containerName="init" Dec 02 01:56:16 crc kubenswrapper[4884]: E1202 01:56:16.988194 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05ad28bd-1166-4977-940f-c0d7a041b177" containerName="glance-log" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.988202 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="05ad28bd-1166-4977-940f-c0d7a041b177" containerName="glance-log" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.988381 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="3058a391-d337-4b74-90c8-9f0d5d81b58e" containerName="glance-log" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.988400 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="05ad28bd-1166-4977-940f-c0d7a041b177" containerName="glance-log" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.988416 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="e47f2b35-d3ff-443a-8f75-37bda0414934" containerName="init" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.988424 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="05ad28bd-1166-4977-940f-c0d7a041b177" containerName="glance-httpd" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.988436 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="3058a391-d337-4b74-90c8-9f0d5d81b58e" containerName="glance-httpd" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.988447 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="b29d1a05-edbf-4146-8c0f-21e846bad95f" containerName="init" Dec 02 01:56:16 crc kubenswrapper[4884]: I1202 01:56:16.989423 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:16.995177 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:16.997775 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.008063 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.008119 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.015534 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.015957 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.016249 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-rvbnc" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.016387 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.016694 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.017076 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.078368 4884 scope.go:117] "RemoveContainer" containerID="e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.136754 4884 scope.go:117] "RemoveContainer" containerID="ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef" Dec 02 01:56:17 crc kubenswrapper[4884]: E1202 01:56:17.137306 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef\": container with ID starting with ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef not found: ID does not exist" containerID="ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.137337 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef"} err="failed to get container status \"ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef\": rpc error: code = NotFound desc = could not find container \"ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef\": container with ID starting with ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef not found: ID does not exist" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.137358 4884 scope.go:117] "RemoveContainer" containerID="e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce" Dec 02 01:56:17 crc kubenswrapper[4884]: E1202 01:56:17.137582 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce\": container with ID starting with e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce not found: ID does not exist" containerID="e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.137622 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce"} err="failed to get container status \"e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce\": rpc error: code = NotFound desc = could not find container \"e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce\": container with ID starting with e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce not found: ID does not exist" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.137639 4884 scope.go:117] "RemoveContainer" containerID="ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.137883 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef"} err="failed to get container status \"ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef\": rpc error: code = NotFound desc = could not find container \"ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef\": container with ID starting with ef5dc8435a9a7d4554d134d6a705a8c1c0424be5df1af4fc99062335dee189ef not found: ID does not exist" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.137901 4884 scope.go:117] "RemoveContainer" containerID="e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.138159 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce"} err="failed to get container status \"e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce\": rpc error: code = NotFound desc = could not find container \"e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce\": container with ID starting with e1e8aadaa77d74490d66910e68e64bfe4a7e686e3a2553435f9a3c15e85c78ce not found: ID does not exist" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171119 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171172 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171231 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx882\" (UniqueName: \"kubernetes.io/projected/0f0a9fed-85dd-4c34-8365-a61044cdbf20-kube-api-access-gx882\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171259 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd7db777-3718-4db3-b144-1fe763e09dae-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171293 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f0a9fed-85dd-4c34-8365-a61044cdbf20-logs\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171313 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171349 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171381 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171395 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdhls\" (UniqueName: \"kubernetes.io/projected/bd7db777-3718-4db3-b144-1fe763e09dae-kube-api-access-jdhls\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171416 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171437 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-scripts\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171463 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171562 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171758 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-config-data\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171791 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f0a9fed-85dd-4c34-8365-a61044cdbf20-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.171848 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd7db777-3718-4db3-b144-1fe763e09dae-logs\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.273483 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.273885 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.273923 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-config-data\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.273941 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f0a9fed-85dd-4c34-8365-a61044cdbf20-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.273959 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd7db777-3718-4db3-b144-1fe763e09dae-logs\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.273979 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.274005 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.274035 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx882\" (UniqueName: \"kubernetes.io/projected/0f0a9fed-85dd-4c34-8365-a61044cdbf20-kube-api-access-gx882\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.274064 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd7db777-3718-4db3-b144-1fe763e09dae-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.274082 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f0a9fed-85dd-4c34-8365-a61044cdbf20-logs\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.274103 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.274141 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.274178 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.274194 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdhls\" (UniqueName: \"kubernetes.io/projected/bd7db777-3718-4db3-b144-1fe763e09dae-kube-api-access-jdhls\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.274218 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.274240 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-scripts\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.277512 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f0a9fed-85dd-4c34-8365-a61044cdbf20-logs\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.278018 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.278911 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd7db777-3718-4db3-b144-1fe763e09dae-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.278985 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f0a9fed-85dd-4c34-8365-a61044cdbf20-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.279101 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.281738 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-scripts\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.283510 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.284584 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-config-data\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.285501 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-scripts\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.285539 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.287529 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd7db777-3718-4db3-b144-1fe763e09dae-logs\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.288527 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.290409 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.291052 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-config-data\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.295044 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx882\" (UniqueName: \"kubernetes.io/projected/0f0a9fed-85dd-4c34-8365-a61044cdbf20-kube-api-access-gx882\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.300532 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdhls\" (UniqueName: \"kubernetes.io/projected/bd7db777-3718-4db3-b144-1fe763e09dae-kube-api-access-jdhls\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.311845 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.318625 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.347968 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.369368 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.388787 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.578970 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-fernet-keys\") pod \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.579105 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-config-data\") pod \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.579176 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-credential-keys\") pod \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.579199 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-scripts\") pod \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.579231 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcghm\" (UniqueName: \"kubernetes.io/projected/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-kube-api-access-gcghm\") pod \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.579267 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-combined-ca-bundle\") pod \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\" (UID: \"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8\") " Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.585925 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-scripts" (OuterVolumeSpecName: "scripts") pod "5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8" (UID: "5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.586725 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-kube-api-access-gcghm" (OuterVolumeSpecName: "kube-api-access-gcghm") pod "5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8" (UID: "5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8"). InnerVolumeSpecName "kube-api-access-gcghm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.588577 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8" (UID: "5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.598221 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8" (UID: "5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.622971 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8" (UID: "5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.631493 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05ad28bd-1166-4977-940f-c0d7a041b177" path="/var/lib/kubelet/pods/05ad28bd-1166-4977-940f-c0d7a041b177/volumes" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.631784 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-config-data" (OuterVolumeSpecName: "config-data") pod "5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8" (UID: "5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.632512 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3058a391-d337-4b74-90c8-9f0d5d81b58e" path="/var/lib/kubelet/pods/3058a391-d337-4b74-90c8-9f0d5d81b58e/volumes" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.683112 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.683538 4884 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.683550 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.683562 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcghm\" (UniqueName: \"kubernetes.io/projected/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-kube-api-access-gcghm\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.683576 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.683585 4884 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.815842 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-pm699" event={"ID":"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8","Type":"ContainerStarted","Data":"3a492fb64d69f7c559b62e62078c74a3c29e4de0c8589a224fff85818c397f94"} Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.819105 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4cqgj" event={"ID":"5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8","Type":"ContainerDied","Data":"8120ba00419ed548c463db81426bcfced683ba2f086808ecc2e1e0adb944eb70"} Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.819147 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8120ba00419ed548c463db81426bcfced683ba2f086808ecc2e1e0adb944eb70" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.819098 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4cqgj" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.840862 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-db-sync-pm699" podStartSLOduration=3.367666494 podStartE2EDuration="31.840840811s" podCreationTimestamp="2025-12-02 01:55:46 +0000 UTC" firstStartedPulling="2025-12-02 01:55:48.209489549 +0000 UTC m=+1164.885326423" lastFinishedPulling="2025-12-02 01:56:16.682663856 +0000 UTC m=+1193.358500740" observedRunningTime="2025-12-02 01:56:17.83271255 +0000 UTC m=+1194.508549454" watchObservedRunningTime="2025-12-02 01:56:17.840840811 +0000 UTC m=+1194.516677695" Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.926287 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-4cqgj"] Dec 02 01:56:17 crc kubenswrapper[4884]: I1202 01:56:17.934129 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-4cqgj"] Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.008297 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-wbxwl"] Dec 02 01:56:18 crc kubenswrapper[4884]: E1202 01:56:18.010813 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8" containerName="keystone-bootstrap" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.010868 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8" containerName="keystone-bootstrap" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.011053 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8" containerName="keystone-bootstrap" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.012217 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.017578 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.017984 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.018094 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.018160 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.018485 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-42pg8" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.020451 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wbxwl"] Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.193360 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-config-data\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.193424 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-combined-ca-bundle\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.193455 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-fernet-keys\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.193823 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-credential-keys\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.193904 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-scripts\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.194097 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f686p\" (UniqueName: \"kubernetes.io/projected/62db24cb-3633-425d-a4b1-f24b4850ce26-kube-api-access-f686p\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.280866 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.289931 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.295287 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f686p\" (UniqueName: \"kubernetes.io/projected/62db24cb-3633-425d-a4b1-f24b4850ce26-kube-api-access-f686p\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.295331 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-config-data\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.295359 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-combined-ca-bundle\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.295378 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-fernet-keys\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.295492 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-credential-keys\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.295520 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-scripts\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.309492 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-config-data\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.310152 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-scripts\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.314505 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-combined-ca-bundle\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.317508 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-fernet-keys\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.321615 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-credential-keys\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.325931 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f686p\" (UniqueName: \"kubernetes.io/projected/62db24cb-3633-425d-a4b1-f24b4850ce26-kube-api-access-f686p\") pod \"keystone-bootstrap-wbxwl\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.343212 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.723859 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d75d4f7c7-492fj"] Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.766642 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-55fc6c9b58-9v749"] Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.771493 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.774937 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.798595 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-55fc6c9b58-9v749"] Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.821714 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.857091 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64d85fdcb9-ncfgl"] Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.873077 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.877468 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-8687f948dd-5zg8m"] Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.879549 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.897767 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.909250 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-horizon-secret-key\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.909294 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59b51ab4-e0c8-42af-92aa-063e66157aa7-logs\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.909331 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsbvn\" (UniqueName: \"kubernetes.io/projected/59b51ab4-e0c8-42af-92aa-063e66157aa7-kube-api-access-zsbvn\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.909380 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/59b51ab4-e0c8-42af-92aa-063e66157aa7-config-data\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.909397 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-horizon-tls-certs\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.909420 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59b51ab4-e0c8-42af-92aa-063e66157aa7-scripts\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.909469 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-combined-ca-bundle\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:18 crc kubenswrapper[4884]: I1202 01:56:18.912853 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8687f948dd-5zg8m"] Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.012808 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsbvn\" (UniqueName: \"kubernetes.io/projected/59b51ab4-e0c8-42af-92aa-063e66157aa7-kube-api-access-zsbvn\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.012929 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/59b51ab4-e0c8-42af-92aa-063e66157aa7-config-data\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.012951 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-horizon-tls-certs\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.012975 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-horizon-tls-certs\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.012995 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59b51ab4-e0c8-42af-92aa-063e66157aa7-scripts\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.013059 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-combined-ca-bundle\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.013078 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-scripts\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.013140 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-logs\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.013158 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szczg\" (UniqueName: \"kubernetes.io/projected/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-kube-api-access-szczg\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.013207 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-horizon-secret-key\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.013233 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-horizon-secret-key\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.013251 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59b51ab4-e0c8-42af-92aa-063e66157aa7-logs\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.013268 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-config-data\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.013285 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-combined-ca-bundle\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.015459 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59b51ab4-e0c8-42af-92aa-063e66157aa7-scripts\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.016226 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59b51ab4-e0c8-42af-92aa-063e66157aa7-logs\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.016974 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/59b51ab4-e0c8-42af-92aa-063e66157aa7-config-data\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.021154 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-horizon-secret-key\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.022479 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-horizon-tls-certs\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.026998 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-combined-ca-bundle\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.033428 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsbvn\" (UniqueName: \"kubernetes.io/projected/59b51ab4-e0c8-42af-92aa-063e66157aa7-kube-api-access-zsbvn\") pod \"horizon-55fc6c9b58-9v749\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.114287 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-horizon-tls-certs\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.114374 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-scripts\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.114424 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-logs\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.114442 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szczg\" (UniqueName: \"kubernetes.io/projected/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-kube-api-access-szczg\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.114466 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-horizon-secret-key\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.114490 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-config-data\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.114508 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-combined-ca-bundle\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.115234 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-logs\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.116792 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-config-data\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.116994 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-scripts\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.119138 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-horizon-tls-certs\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.119314 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-horizon-secret-key\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.119879 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-combined-ca-bundle\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.134993 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.138345 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szczg\" (UniqueName: \"kubernetes.io/projected/19afec6c-37bd-4e1a-9f96-c7e37c1e7748-kube-api-access-szczg\") pod \"horizon-8687f948dd-5zg8m\" (UID: \"19afec6c-37bd-4e1a-9f96-c7e37c1e7748\") " pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.198175 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:19 crc kubenswrapper[4884]: I1202 01:56:19.625559 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8" path="/var/lib/kubelet/pods/5e1fc554-ace1-4c3d-ac7f-5af6b9e932a8/volumes" Dec 02 01:56:22 crc kubenswrapper[4884]: I1202 01:56:22.896291 4884 generic.go:334] "Generic (PLEG): container finished" podID="5ba1201c-49ba-4c31-bf2a-8069ea8a16c8" containerID="3a492fb64d69f7c559b62e62078c74a3c29e4de0c8589a224fff85818c397f94" exitCode=0 Dec 02 01:56:22 crc kubenswrapper[4884]: I1202 01:56:22.896376 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-pm699" event={"ID":"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8","Type":"ContainerDied","Data":"3a492fb64d69f7c559b62e62078c74a3c29e4de0c8589a224fff85818c397f94"} Dec 02 01:56:23 crc kubenswrapper[4884]: I1202 01:56:23.402894 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:23 crc kubenswrapper[4884]: I1202 01:56:23.478043 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-5kq68"] Dec 02 01:56:23 crc kubenswrapper[4884]: I1202 01:56:23.478304 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-5kq68" podUID="ff7b03af-7163-47fa-9397-ce25148d27c5" containerName="dnsmasq-dns" containerID="cri-o://1be5bbc237238ae274ed7f2a27e79d37fddd40ccdf473e477fc749394117ff42" gracePeriod=10 Dec 02 01:56:23 crc kubenswrapper[4884]: I1202 01:56:23.908329 4884 generic.go:334] "Generic (PLEG): container finished" podID="ff7b03af-7163-47fa-9397-ce25148d27c5" containerID="1be5bbc237238ae274ed7f2a27e79d37fddd40ccdf473e477fc749394117ff42" exitCode=0 Dec 02 01:56:23 crc kubenswrapper[4884]: I1202 01:56:23.908885 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-5kq68" event={"ID":"ff7b03af-7163-47fa-9397-ce25148d27c5","Type":"ContainerDied","Data":"1be5bbc237238ae274ed7f2a27e79d37fddd40ccdf473e477fc749394117ff42"} Dec 02 01:56:30 crc kubenswrapper[4884]: E1202 01:56:30.735696 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 02 01:56:30 crc kubenswrapper[4884]: E1202 01:56:30.736493 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n664h69h64ch68ch648h669h57dh5b4hcch67fh5b4hch565h55bhbh5bdh5c5h57h589h555h66dh5b5h5d8h5fbhc8h58bh5c7hdbh559h5d5h68fhbbq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sp5ss,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-b6f66b857-h6k98_openstack(d0550743-d9cf-4b93-a2df-86bd40d2dd3d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 01:56:30 crc kubenswrapper[4884]: E1202 01:56:30.746036 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-b6f66b857-h6k98" podUID="d0550743-d9cf-4b93-a2df-86bd40d2dd3d" Dec 02 01:56:32 crc kubenswrapper[4884]: I1202 01:56:32.010530 4884 generic.go:334] "Generic (PLEG): container finished" podID="d95342e1-6aff-4feb-9309-26ffa69e786b" containerID="a5d68bf6080508333b0425d2b369cbf613c67a868be4a5af9d924e71ccbbd979" exitCode=0 Dec 02 01:56:32 crc kubenswrapper[4884]: I1202 01:56:32.010721 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jn9fg" event={"ID":"d95342e1-6aff-4feb-9309-26ffa69e786b","Type":"ContainerDied","Data":"a5d68bf6080508333b0425d2b369cbf613c67a868be4a5af9d924e71ccbbd979"} Dec 02 01:56:32 crc kubenswrapper[4884]: I1202 01:56:32.202182 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-5kq68" podUID="ff7b03af-7163-47fa-9397-ce25148d27c5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.119:5353: i/o timeout" Dec 02 01:56:37 crc kubenswrapper[4884]: I1202 01:56:37.202986 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-5kq68" podUID="ff7b03af-7163-47fa-9397-ce25148d27c5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.119:5353: i/o timeout" Dec 02 01:56:38 crc kubenswrapper[4884]: E1202 01:56:38.063020 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 02 01:56:38 crc kubenswrapper[4884]: E1202 01:56:38.063564 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n64h55ch55bh5dbh8fh96h5fdh54chf9h564h646h8ch5f8h664h98hbbh559h5f6h69h5h57bh68dh674h668hc9h58fh66dh688h5cch66fh664h577q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jd55b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-64d85fdcb9-ncfgl_openstack(8f9d2899-4e00-467e-afbe-214f40042543): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 01:56:38 crc kubenswrapper[4884]: E1202 01:56:38.067496 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-64d85fdcb9-ncfgl" podUID="8f9d2899-4e00-467e-afbe-214f40042543" Dec 02 01:56:38 crc kubenswrapper[4884]: E1202 01:56:38.083411 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 02 01:56:38 crc kubenswrapper[4884]: E1202 01:56:38.083598 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5b5h55fh6ch5c8h699h688hf6h66bh66ch669h5d6hcbh554h5c5h669h645h5cdh594h655h8bh6dhc6h697h8ch5f8h9ch56ch56dh57bh7fh96h696q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:yes,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dzjts,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7d75d4f7c7-492fj_openstack(ea9acb40-6e2e-4873-9a88-66fb8830d688): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 01:56:38 crc kubenswrapper[4884]: E1202 01:56:38.086813 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7d75d4f7c7-492fj" podUID="ea9acb40-6e2e-4873-9a88-66fb8830d688" Dec 02 01:56:38 crc kubenswrapper[4884]: E1202 01:56:38.618776 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 02 01:56:38 crc kubenswrapper[4884]: E1202 01:56:38.618992 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pjdq9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-wmhnq_openstack(fe8c221e-e320-4358-b940-436f38a423d3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 01:56:38 crc kubenswrapper[4884]: E1202 01:56:38.620225 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-wmhnq" podUID="fe8c221e-e320-4358-b940-436f38a423d3" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.738233 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-pm699" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.744002 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.748929 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.762447 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jn9fg" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.829472 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-combined-ca-bundle\") pod \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.829790 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-dns-svc\") pod \"ff7b03af-7163-47fa-9397-ce25148d27c5\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.829815 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-ovsdbserver-nb\") pod \"ff7b03af-7163-47fa-9397-ce25148d27c5\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.829862 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh94v\" (UniqueName: \"kubernetes.io/projected/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-kube-api-access-lh94v\") pod \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.829878 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d95342e1-6aff-4feb-9309-26ffa69e786b-config\") pod \"d95342e1-6aff-4feb-9309-26ffa69e786b\" (UID: \"d95342e1-6aff-4feb-9309-26ffa69e786b\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.829897 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55k9t\" (UniqueName: \"kubernetes.io/projected/ff7b03af-7163-47fa-9397-ce25148d27c5-kube-api-access-55k9t\") pod \"ff7b03af-7163-47fa-9397-ce25148d27c5\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.829925 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-config-data\") pod \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.829945 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-config\") pod \"ff7b03af-7163-47fa-9397-ce25148d27c5\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.829991 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-horizon-secret-key\") pod \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.830017 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-db-sync-config-data\") pod \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.830035 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-logs\") pod \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.830051 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-config-data\") pod \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\" (UID: \"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.830110 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-ovsdbserver-sb\") pod \"ff7b03af-7163-47fa-9397-ce25148d27c5\" (UID: \"ff7b03af-7163-47fa-9397-ce25148d27c5\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.830143 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp5ss\" (UniqueName: \"kubernetes.io/projected/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-kube-api-access-sp5ss\") pod \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.830160 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmk6l\" (UniqueName: \"kubernetes.io/projected/d95342e1-6aff-4feb-9309-26ffa69e786b-kube-api-access-mmk6l\") pod \"d95342e1-6aff-4feb-9309-26ffa69e786b\" (UID: \"d95342e1-6aff-4feb-9309-26ffa69e786b\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.830194 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-scripts\") pod \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\" (UID: \"d0550743-d9cf-4b93-a2df-86bd40d2dd3d\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.830237 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95342e1-6aff-4feb-9309-26ffa69e786b-combined-ca-bundle\") pod \"d95342e1-6aff-4feb-9309-26ffa69e786b\" (UID: \"d95342e1-6aff-4feb-9309-26ffa69e786b\") " Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.836397 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-kube-api-access-lh94v" (OuterVolumeSpecName: "kube-api-access-lh94v") pod "5ba1201c-49ba-4c31-bf2a-8069ea8a16c8" (UID: "5ba1201c-49ba-4c31-bf2a-8069ea8a16c8"). InnerVolumeSpecName "kube-api-access-lh94v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.836924 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-logs" (OuterVolumeSpecName: "logs") pod "d0550743-d9cf-4b93-a2df-86bd40d2dd3d" (UID: "d0550743-d9cf-4b93-a2df-86bd40d2dd3d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.838485 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5ba1201c-49ba-4c31-bf2a-8069ea8a16c8" (UID: "5ba1201c-49ba-4c31-bf2a-8069ea8a16c8"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.844061 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-scripts" (OuterVolumeSpecName: "scripts") pod "d0550743-d9cf-4b93-a2df-86bd40d2dd3d" (UID: "d0550743-d9cf-4b93-a2df-86bd40d2dd3d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.845195 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-config-data" (OuterVolumeSpecName: "config-data") pod "d0550743-d9cf-4b93-a2df-86bd40d2dd3d" (UID: "d0550743-d9cf-4b93-a2df-86bd40d2dd3d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.845303 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d0550743-d9cf-4b93-a2df-86bd40d2dd3d" (UID: "d0550743-d9cf-4b93-a2df-86bd40d2dd3d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.851961 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff7b03af-7163-47fa-9397-ce25148d27c5-kube-api-access-55k9t" (OuterVolumeSpecName: "kube-api-access-55k9t") pod "ff7b03af-7163-47fa-9397-ce25148d27c5" (UID: "ff7b03af-7163-47fa-9397-ce25148d27c5"). InnerVolumeSpecName "kube-api-access-55k9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.853042 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-kube-api-access-sp5ss" (OuterVolumeSpecName: "kube-api-access-sp5ss") pod "d0550743-d9cf-4b93-a2df-86bd40d2dd3d" (UID: "d0550743-d9cf-4b93-a2df-86bd40d2dd3d"). InnerVolumeSpecName "kube-api-access-sp5ss". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.854791 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d95342e1-6aff-4feb-9309-26ffa69e786b-kube-api-access-mmk6l" (OuterVolumeSpecName: "kube-api-access-mmk6l") pod "d95342e1-6aff-4feb-9309-26ffa69e786b" (UID: "d95342e1-6aff-4feb-9309-26ffa69e786b"). InnerVolumeSpecName "kube-api-access-mmk6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.897765 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95342e1-6aff-4feb-9309-26ffa69e786b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d95342e1-6aff-4feb-9309-26ffa69e786b" (UID: "d95342e1-6aff-4feb-9309-26ffa69e786b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.915206 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ba1201c-49ba-4c31-bf2a-8069ea8a16c8" (UID: "5ba1201c-49ba-4c31-bf2a-8069ea8a16c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.918589 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95342e1-6aff-4feb-9309-26ffa69e786b-config" (OuterVolumeSpecName: "config") pod "d95342e1-6aff-4feb-9309-26ffa69e786b" (UID: "d95342e1-6aff-4feb-9309-26ffa69e786b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.920187 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ff7b03af-7163-47fa-9397-ce25148d27c5" (UID: "ff7b03af-7163-47fa-9397-ce25148d27c5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.932398 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmk6l\" (UniqueName: \"kubernetes.io/projected/d95342e1-6aff-4feb-9309-26ffa69e786b-kube-api-access-mmk6l\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.932429 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.932438 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95342e1-6aff-4feb-9309-26ffa69e786b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.932448 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.932456 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.932465 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh94v\" (UniqueName: \"kubernetes.io/projected/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-kube-api-access-lh94v\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.932473 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d95342e1-6aff-4feb-9309-26ffa69e786b-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.932481 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55k9t\" (UniqueName: \"kubernetes.io/projected/ff7b03af-7163-47fa-9397-ce25148d27c5-kube-api-access-55k9t\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.932489 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.932498 4884 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.932506 4884 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.932515 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.932527 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp5ss\" (UniqueName: \"kubernetes.io/projected/d0550743-d9cf-4b93-a2df-86bd40d2dd3d-kube-api-access-sp5ss\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.935389 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ff7b03af-7163-47fa-9397-ce25148d27c5" (UID: "ff7b03af-7163-47fa-9397-ce25148d27c5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.936600 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-config-data" (OuterVolumeSpecName: "config-data") pod "5ba1201c-49ba-4c31-bf2a-8069ea8a16c8" (UID: "5ba1201c-49ba-4c31-bf2a-8069ea8a16c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.951554 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ff7b03af-7163-47fa-9397-ce25148d27c5" (UID: "ff7b03af-7163-47fa-9397-ce25148d27c5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:38 crc kubenswrapper[4884]: I1202 01:56:38.953080 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-config" (OuterVolumeSpecName: "config") pod "ff7b03af-7163-47fa-9397-ce25148d27c5" (UID: "ff7b03af-7163-47fa-9397-ce25148d27c5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.034539 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.034573 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.034583 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.034591 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff7b03af-7163-47fa-9397-ce25148d27c5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.103044 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-b6f66b857-h6k98" event={"ID":"d0550743-d9cf-4b93-a2df-86bd40d2dd3d","Type":"ContainerDied","Data":"6c87bb72dcade19b2a9a857c8cc38b0e51b4b38933fc5f44e251dd9b713657a6"} Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.103068 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-b6f66b857-h6k98" Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.106794 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jn9fg" Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.106827 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jn9fg" event={"ID":"d95342e1-6aff-4feb-9309-26ffa69e786b","Type":"ContainerDied","Data":"4f1e5c894cb8298a0eae5ae62b12d4575ba875e1b943cac586ece27c6ea21f73"} Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.106878 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f1e5c894cb8298a0eae5ae62b12d4575ba875e1b943cac586ece27c6ea21f73" Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.109098 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-5kq68" event={"ID":"ff7b03af-7163-47fa-9397-ce25148d27c5","Type":"ContainerDied","Data":"50515de5c8fd063aa9355aeab5d9f9cbabe5330ef6e86b029d20da879387821e"} Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.109142 4884 scope.go:117] "RemoveContainer" containerID="1be5bbc237238ae274ed7f2a27e79d37fddd40ccdf473e477fc749394117ff42" Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.109254 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-5kq68" Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.118066 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-db-sync-pm699" event={"ID":"5ba1201c-49ba-4c31-bf2a-8069ea8a16c8","Type":"ContainerDied","Data":"f46a8eaec2f4213573dff68f66db7462342045321d98f1790820d49739ce6d94"} Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.118130 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f46a8eaec2f4213573dff68f66db7462342045321d98f1790820d49739ce6d94" Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.118312 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-db-sync-pm699" Dec 02 01:56:39 crc kubenswrapper[4884]: E1202 01:56:39.120914 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-wmhnq" podUID="fe8c221e-e320-4358-b940-436f38a423d3" Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.164822 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-5kq68"] Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.191418 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-5kq68"] Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.256052 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-b6f66b857-h6k98"] Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.262832 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-b6f66b857-h6k98"] Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.632065 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0550743-d9cf-4b93-a2df-86bd40d2dd3d" path="/var/lib/kubelet/pods/d0550743-d9cf-4b93-a2df-86bd40d2dd3d/volumes" Dec 02 01:56:39 crc kubenswrapper[4884]: I1202 01:56:39.633363 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff7b03af-7163-47fa-9397-ce25148d27c5" path="/var/lib/kubelet/pods/ff7b03af-7163-47fa-9397-ce25148d27c5/volumes" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.063384 4884 scope.go:117] "RemoveContainer" containerID="4124739ea4376e86d74c8f412d23ddbc69be6289c95f822b19b0763e879533b2" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.063562 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-whpnl"] Dec 02 01:56:40 crc kubenswrapper[4884]: E1202 01:56:40.063966 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95342e1-6aff-4feb-9309-26ffa69e786b" containerName="neutron-db-sync" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.064038 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95342e1-6aff-4feb-9309-26ffa69e786b" containerName="neutron-db-sync" Dec 02 01:56:40 crc kubenswrapper[4884]: E1202 01:56:40.064090 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7b03af-7163-47fa-9397-ce25148d27c5" containerName="init" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.064135 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7b03af-7163-47fa-9397-ce25148d27c5" containerName="init" Dec 02 01:56:40 crc kubenswrapper[4884]: E1202 01:56:40.064205 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7b03af-7163-47fa-9397-ce25148d27c5" containerName="dnsmasq-dns" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.064250 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7b03af-7163-47fa-9397-ce25148d27c5" containerName="dnsmasq-dns" Dec 02 01:56:40 crc kubenswrapper[4884]: E1202 01:56:40.064303 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ba1201c-49ba-4c31-bf2a-8069ea8a16c8" containerName="watcher-db-sync" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.064349 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ba1201c-49ba-4c31-bf2a-8069ea8a16c8" containerName="watcher-db-sync" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.064559 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff7b03af-7163-47fa-9397-ce25148d27c5" containerName="dnsmasq-dns" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.064617 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95342e1-6aff-4feb-9309-26ffa69e786b" containerName="neutron-db-sync" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.064680 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ba1201c-49ba-4c31-bf2a-8069ea8a16c8" containerName="watcher-db-sync" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.065690 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: E1202 01:56:40.078395 4884 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 02 01:56:40 crc kubenswrapper[4884]: E1202 01:56:40.078597 4884 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wd49x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-c9rsj_openstack(f0d68bd4-6c59-4e41-b32e-f7918d3b4ada): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 02 01:56:40 crc kubenswrapper[4884]: E1202 01:56:40.079668 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-c9rsj" podUID="f0d68bd4-6c59-4e41-b32e-f7918d3b4ada" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.083428 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-whpnl"] Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.164078 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-config\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.164202 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.164234 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.164255 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-dns-svc\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.164272 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbfn7\" (UniqueName: \"kubernetes.io/projected/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-kube-api-access-qbfn7\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.164302 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.209009 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.210170 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d75d4f7c7-492fj" event={"ID":"ea9acb40-6e2e-4873-9a88-66fb8830d688","Type":"ContainerDied","Data":"043624d531af668431812d880ea3320f4866dd8182efabc11e517cbe5c9452be"} Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.210220 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="043624d531af668431812d880ea3320f4866dd8182efabc11e517cbe5c9452be" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.219489 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.220688 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.227695 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.229052 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-watcher-dockercfg-r5l44" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.246365 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64d85fdcb9-ncfgl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.249812 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64d85fdcb9-ncfgl" event={"ID":"8f9d2899-4e00-467e-afbe-214f40042543","Type":"ContainerDied","Data":"e3e9b26d58145e3c24fa719a8445d86e9ff72fd967c9ee1adb0c73b71391440e"} Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.265327 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f9d2899-4e00-467e-afbe-214f40042543-horizon-secret-key\") pod \"8f9d2899-4e00-467e-afbe-214f40042543\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.265423 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f9d2899-4e00-467e-afbe-214f40042543-config-data\") pod \"8f9d2899-4e00-467e-afbe-214f40042543\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.265472 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f9d2899-4e00-467e-afbe-214f40042543-scripts\") pod \"8f9d2899-4e00-467e-afbe-214f40042543\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.265577 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd55b\" (UniqueName: \"kubernetes.io/projected/8f9d2899-4e00-467e-afbe-214f40042543-kube-api-access-jd55b\") pod \"8f9d2899-4e00-467e-afbe-214f40042543\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.265660 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f9d2899-4e00-467e-afbe-214f40042543-logs\") pod \"8f9d2899-4e00-467e-afbe-214f40042543\" (UID: \"8f9d2899-4e00-467e-afbe-214f40042543\") " Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.265880 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-config\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.265994 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.266026 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.266048 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-dns-svc\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.266064 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbfn7\" (UniqueName: \"kubernetes.io/projected/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-kube-api-access-qbfn7\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.266097 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.271713 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-ovsdbserver-sb\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.280802 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f9d2899-4e00-467e-afbe-214f40042543-scripts" (OuterVolumeSpecName: "scripts") pod "8f9d2899-4e00-467e-afbe-214f40042543" (UID: "8f9d2899-4e00-467e-afbe-214f40042543"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.281261 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f9d2899-4e00-467e-afbe-214f40042543-config-data" (OuterVolumeSpecName: "config-data") pod "8f9d2899-4e00-467e-afbe-214f40042543" (UID: "8f9d2899-4e00-467e-afbe-214f40042543"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.281816 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f9d2899-4e00-467e-afbe-214f40042543-logs" (OuterVolumeSpecName: "logs") pod "8f9d2899-4e00-467e-afbe-214f40042543" (UID: "8f9d2899-4e00-467e-afbe-214f40042543"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.289979 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-dns-svc\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.290103 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.290141 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-ovsdbserver-nb\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.290255 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-dns-swift-storage-0\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.295247 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-config\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.295377 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f9d2899-4e00-467e-afbe-214f40042543-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8f9d2899-4e00-467e-afbe-214f40042543" (UID: "8f9d2899-4e00-467e-afbe-214f40042543"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.304008 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f9d2899-4e00-467e-afbe-214f40042543-kube-api-access-jd55b" (OuterVolumeSpecName: "kube-api-access-jd55b") pod "8f9d2899-4e00-467e-afbe-214f40042543" (UID: "8f9d2899-4e00-467e-afbe-214f40042543"). InnerVolumeSpecName "kube-api-access-jd55b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.306447 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.311862 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbfn7\" (UniqueName: \"kubernetes.io/projected/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-kube-api-access-qbfn7\") pod \"dnsmasq-dns-55f844cf75-whpnl\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.315559 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-applier-0"] Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.319837 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.320914 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.322441 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-applier-config-data" Dec 02 01:56:40 crc kubenswrapper[4884]: E1202 01:56:40.324032 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-c9rsj" podUID="f0d68bd4-6c59-4e41-b32e-f7918d3b4ada" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.337323 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.339599 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.341178 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.352403 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.369189 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzjts\" (UniqueName: \"kubernetes.io/projected/ea9acb40-6e2e-4873-9a88-66fb8830d688-kube-api-access-dzjts\") pod \"ea9acb40-6e2e-4873-9a88-66fb8830d688\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.369283 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ea9acb40-6e2e-4873-9a88-66fb8830d688-horizon-secret-key\") pod \"ea9acb40-6e2e-4873-9a88-66fb8830d688\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.369369 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea9acb40-6e2e-4873-9a88-66fb8830d688-config-data\") pod \"ea9acb40-6e2e-4873-9a88-66fb8830d688\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.369486 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea9acb40-6e2e-4873-9a88-66fb8830d688-scripts\") pod \"ea9acb40-6e2e-4873-9a88-66fb8830d688\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.369528 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea9acb40-6e2e-4873-9a88-66fb8830d688-logs\") pod \"ea9acb40-6e2e-4873-9a88-66fb8830d688\" (UID: \"ea9acb40-6e2e-4873-9a88-66fb8830d688\") " Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.369731 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2417b2c6-1407-4e68-aed3-b4d14802bc9e-logs\") pod \"watcher-decision-engine-0\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.369851 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-config-data\") pod \"watcher-decision-engine-0\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.369926 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.370017 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.370065 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s47z\" (UniqueName: \"kubernetes.io/projected/2417b2c6-1407-4e68-aed3-b4d14802bc9e-kube-api-access-2s47z\") pod \"watcher-decision-engine-0\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.370151 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd55b\" (UniqueName: \"kubernetes.io/projected/8f9d2899-4e00-467e-afbe-214f40042543-kube-api-access-jd55b\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.370161 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f9d2899-4e00-467e-afbe-214f40042543-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.370170 4884 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8f9d2899-4e00-467e-afbe-214f40042543-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.370179 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8f9d2899-4e00-467e-afbe-214f40042543-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.370191 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8f9d2899-4e00-467e-afbe-214f40042543-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.373655 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea9acb40-6e2e-4873-9a88-66fb8830d688-logs" (OuterVolumeSpecName: "logs") pod "ea9acb40-6e2e-4873-9a88-66fb8830d688" (UID: "ea9acb40-6e2e-4873-9a88-66fb8830d688"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.374082 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea9acb40-6e2e-4873-9a88-66fb8830d688-config-data" (OuterVolumeSpecName: "config-data") pod "ea9acb40-6e2e-4873-9a88-66fb8830d688" (UID: "ea9acb40-6e2e-4873-9a88-66fb8830d688"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.374567 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea9acb40-6e2e-4873-9a88-66fb8830d688-scripts" (OuterVolumeSpecName: "scripts") pod "ea9acb40-6e2e-4873-9a88-66fb8830d688" (UID: "ea9acb40-6e2e-4873-9a88-66fb8830d688"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.376531 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea9acb40-6e2e-4873-9a88-66fb8830d688-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ea9acb40-6e2e-4873-9a88-66fb8830d688" (UID: "ea9acb40-6e2e-4873-9a88-66fb8830d688"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.382025 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea9acb40-6e2e-4873-9a88-66fb8830d688-kube-api-access-dzjts" (OuterVolumeSpecName: "kube-api-access-dzjts") pod "ea9acb40-6e2e-4873-9a88-66fb8830d688" (UID: "ea9acb40-6e2e-4873-9a88-66fb8830d688"). InnerVolumeSpecName "kube-api-access-dzjts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.460236 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.472630 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.472699 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25acc2a6-f07c-44c1-aba6-b5f2e38d39ce-config-data\") pod \"watcher-applier-0\" (UID: \"25acc2a6-f07c-44c1-aba6-b5f2e38d39ce\") " pod="openstack/watcher-applier-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.472736 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2417b2c6-1407-4e68-aed3-b4d14802bc9e-logs\") pod \"watcher-decision-engine-0\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.472851 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4cjb\" (UniqueName: \"kubernetes.io/projected/25acc2a6-f07c-44c1-aba6-b5f2e38d39ce-kube-api-access-q4cjb\") pod \"watcher-applier-0\" (UID: \"25acc2a6-f07c-44c1-aba6-b5f2e38d39ce\") " pod="openstack/watcher-applier-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.472916 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-config-data\") pod \"watcher-decision-engine-0\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.473041 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-logs\") pod \"watcher-api-0\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.473076 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.473105 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.473154 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25acc2a6-f07c-44c1-aba6-b5f2e38d39ce-logs\") pod \"watcher-applier-0\" (UID: \"25acc2a6-f07c-44c1-aba6-b5f2e38d39ce\") " pod="openstack/watcher-applier-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.473198 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlzhl\" (UniqueName: \"kubernetes.io/projected/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-kube-api-access-wlzhl\") pod \"watcher-api-0\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.473199 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2417b2c6-1407-4e68-aed3-b4d14802bc9e-logs\") pod \"watcher-decision-engine-0\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.473269 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.473349 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-config-data\") pod \"watcher-api-0\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.473383 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s47z\" (UniqueName: \"kubernetes.io/projected/2417b2c6-1407-4e68-aed3-b4d14802bc9e-kube-api-access-2s47z\") pod \"watcher-decision-engine-0\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.473405 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25acc2a6-f07c-44c1-aba6-b5f2e38d39ce-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"25acc2a6-f07c-44c1-aba6-b5f2e38d39ce\") " pod="openstack/watcher-applier-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.473603 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ea9acb40-6e2e-4873-9a88-66fb8830d688-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.473623 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea9acb40-6e2e-4873-9a88-66fb8830d688-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.473634 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea9acb40-6e2e-4873-9a88-66fb8830d688-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.473645 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzjts\" (UniqueName: \"kubernetes.io/projected/ea9acb40-6e2e-4873-9a88-66fb8830d688-kube-api-access-dzjts\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.473655 4884 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ea9acb40-6e2e-4873-9a88-66fb8830d688-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.476560 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-config-data\") pod \"watcher-decision-engine-0\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.477010 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.478961 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.489608 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s47z\" (UniqueName: \"kubernetes.io/projected/2417b2c6-1407-4e68-aed3-b4d14802bc9e-kube-api-access-2s47z\") pod \"watcher-decision-engine-0\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.493239 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-679d6c649b-jzn47"] Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.494855 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.497402 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8bkvq" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.497671 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.497992 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.504180 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.510401 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-679d6c649b-jzn47"] Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.589936 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-httpd-config\") pod \"neutron-679d6c649b-jzn47\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.590209 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4cjb\" (UniqueName: \"kubernetes.io/projected/25acc2a6-f07c-44c1-aba6-b5f2e38d39ce-kube-api-access-q4cjb\") pod \"watcher-applier-0\" (UID: \"25acc2a6-f07c-44c1-aba6-b5f2e38d39ce\") " pod="openstack/watcher-applier-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.590228 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-combined-ca-bundle\") pod \"neutron-679d6c649b-jzn47\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.590354 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-config\") pod \"neutron-679d6c649b-jzn47\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.590404 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-logs\") pod \"watcher-api-0\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.590428 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmzzz\" (UniqueName: \"kubernetes.io/projected/321c69df-0c85-4378-af0f-2b127be9b6a9-kube-api-access-bmzzz\") pod \"neutron-679d6c649b-jzn47\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.590509 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.590575 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25acc2a6-f07c-44c1-aba6-b5f2e38d39ce-logs\") pod \"watcher-applier-0\" (UID: \"25acc2a6-f07c-44c1-aba6-b5f2e38d39ce\") " pod="openstack/watcher-applier-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.591447 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-logs\") pod \"watcher-api-0\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.590610 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlzhl\" (UniqueName: \"kubernetes.io/projected/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-kube-api-access-wlzhl\") pod \"watcher-api-0\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.593020 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-config-data\") pod \"watcher-api-0\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.593054 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25acc2a6-f07c-44c1-aba6-b5f2e38d39ce-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"25acc2a6-f07c-44c1-aba6-b5f2e38d39ce\") " pod="openstack/watcher-applier-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.593402 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25acc2a6-f07c-44c1-aba6-b5f2e38d39ce-logs\") pod \"watcher-applier-0\" (UID: \"25acc2a6-f07c-44c1-aba6-b5f2e38d39ce\") " pod="openstack/watcher-applier-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.602565 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.602625 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25acc2a6-f07c-44c1-aba6-b5f2e38d39ce-config-data\") pod \"watcher-applier-0\" (UID: \"25acc2a6-f07c-44c1-aba6-b5f2e38d39ce\") " pod="openstack/watcher-applier-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.602673 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-ovndb-tls-certs\") pod \"neutron-679d6c649b-jzn47\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.607890 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.615882 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25acc2a6-f07c-44c1-aba6-b5f2e38d39ce-config-data\") pod \"watcher-applier-0\" (UID: \"25acc2a6-f07c-44c1-aba6-b5f2e38d39ce\") " pod="openstack/watcher-applier-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.621410 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25acc2a6-f07c-44c1-aba6-b5f2e38d39ce-combined-ca-bundle\") pod \"watcher-applier-0\" (UID: \"25acc2a6-f07c-44c1-aba6-b5f2e38d39ce\") " pod="openstack/watcher-applier-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.622937 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlzhl\" (UniqueName: \"kubernetes.io/projected/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-kube-api-access-wlzhl\") pod \"watcher-api-0\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.627117 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-config-data\") pod \"watcher-api-0\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.630374 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4cjb\" (UniqueName: \"kubernetes.io/projected/25acc2a6-f07c-44c1-aba6-b5f2e38d39ce-kube-api-access-q4cjb\") pod \"watcher-applier-0\" (UID: \"25acc2a6-f07c-44c1-aba6-b5f2e38d39ce\") " pod="openstack/watcher-applier-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.641977 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.660837 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64d85fdcb9-ncfgl"] Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.666805 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-64d85fdcb9-ncfgl"] Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.688474 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.718776 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.719298 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-applier-0" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.722622 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-ovndb-tls-certs\") pod \"neutron-679d6c649b-jzn47\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.722668 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-httpd-config\") pod \"neutron-679d6c649b-jzn47\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.722694 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-combined-ca-bundle\") pod \"neutron-679d6c649b-jzn47\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.722812 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-config\") pod \"neutron-679d6c649b-jzn47\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.722839 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmzzz\" (UniqueName: \"kubernetes.io/projected/321c69df-0c85-4378-af0f-2b127be9b6a9-kube-api-access-bmzzz\") pod \"neutron-679d6c649b-jzn47\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.730326 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-ovndb-tls-certs\") pod \"neutron-679d6c649b-jzn47\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.730955 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-combined-ca-bundle\") pod \"neutron-679d6c649b-jzn47\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.730970 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-httpd-config\") pod \"neutron-679d6c649b-jzn47\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.735522 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-config\") pod \"neutron-679d6c649b-jzn47\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.748141 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmzzz\" (UniqueName: \"kubernetes.io/projected/321c69df-0c85-4378-af0f-2b127be9b6a9-kube-api-access-bmzzz\") pod \"neutron-679d6c649b-jzn47\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:40 crc kubenswrapper[4884]: I1202 01:56:40.834967 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.044381 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-55fc6c9b58-9v749"] Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.072170 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8687f948dd-5zg8m"] Dec 02 01:56:41 crc kubenswrapper[4884]: W1202 01:56:41.082701 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19afec6c_37bd_4e1a_9f96_c7e37c1e7748.slice/crio-d692b9b28bb2e5e7e2ee69c2d91449a20b6c51a185df3eacd51fc7d1e808fce3 WatchSource:0}: Error finding container d692b9b28bb2e5e7e2ee69c2d91449a20b6c51a185df3eacd51fc7d1e808fce3: Status 404 returned error can't find the container with id d692b9b28bb2e5e7e2ee69c2d91449a20b6c51a185df3eacd51fc7d1e808fce3 Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.089248 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-wbxwl"] Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.115415 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.140578 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:56:41 crc kubenswrapper[4884]: W1202 01:56:41.142183 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd7db777_3718_4db3_b144_1fe763e09dae.slice/crio-aaed6e805751b822c9c9f96bfff8ff9a0ec8cde86e69d3256920878330d61bc0 WatchSource:0}: Error finding container aaed6e805751b822c9c9f96bfff8ff9a0ec8cde86e69d3256920878330d61bc0: Status 404 returned error can't find the container with id aaed6e805751b822c9c9f96bfff8ff9a0ec8cde86e69d3256920878330d61bc0 Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.173218 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-whpnl"] Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.296309 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55fc6c9b58-9v749" event={"ID":"59b51ab4-e0c8-42af-92aa-063e66157aa7","Type":"ContainerStarted","Data":"d4f7cd50430eda19403e476a0d7de9664b4688b9b16b133b01d024e84514f2b5"} Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.299721 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.300907 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wbxwl" event={"ID":"62db24cb-3633-425d-a4b1-f24b4850ce26","Type":"ContainerStarted","Data":"a36efa1a0b410641f38a458cf14451ad053341e02c79d177546d9af853567539"} Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.303365 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bd7db777-3718-4db3-b144-1fe763e09dae","Type":"ContainerStarted","Data":"aaed6e805751b822c9c9f96bfff8ff9a0ec8cde86e69d3256920878330d61bc0"} Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.314144 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5vj4w" event={"ID":"37412014-287b-42f1-9437-ad164aa34f52","Type":"ContainerStarted","Data":"decead557aa4909bdaee4b0c705b59f7c4c2cfb3584978e712a87366ff7ab8ac"} Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.321240 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8687f948dd-5zg8m" event={"ID":"19afec6c-37bd-4e1a-9f96-c7e37c1e7748","Type":"ContainerStarted","Data":"d692b9b28bb2e5e7e2ee69c2d91449a20b6c51a185df3eacd51fc7d1e808fce3"} Dec 02 01:56:41 crc kubenswrapper[4884]: W1202 01:56:41.325009 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2417b2c6_1407_4e68_aed3_b4d14802bc9e.slice/crio-0abe34c3f054d82b6f308f9fd37d9d645be0e0251e0dcf4d517b750683040533 WatchSource:0}: Error finding container 0abe34c3f054d82b6f308f9fd37d9d645be0e0251e0dcf4d517b750683040533: Status 404 returned error can't find the container with id 0abe34c3f054d82b6f308f9fd37d9d645be0e0251e0dcf4d517b750683040533 Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.328882 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-5vj4w" podStartSLOduration=3.150795269 podStartE2EDuration="31.328863244s" podCreationTimestamp="2025-12-02 01:56:10 +0000 UTC" firstStartedPulling="2025-12-02 01:56:11.776908019 +0000 UTC m=+1188.452744893" lastFinishedPulling="2025-12-02 01:56:39.954975974 +0000 UTC m=+1216.630812868" observedRunningTime="2025-12-02 01:56:41.325308586 +0000 UTC m=+1218.001145470" watchObservedRunningTime="2025-12-02 01:56:41.328863244 +0000 UTC m=+1218.004700128" Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.352428 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45","Type":"ContainerStarted","Data":"b40142baf48473a349fd9c65a988c97d8d72dedfe8f3737777935794684cb579"} Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.355943 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-whpnl" event={"ID":"0c25ce5d-09d7-4a8b-be58-baa2d27191e8","Type":"ContainerStarted","Data":"6a769b7aa74d7a6f402a4e164e36bde4ed1dd57ea2a2ff87d9c12a5c2599098b"} Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.355975 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d75d4f7c7-492fj" Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.402286 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-applier-0"] Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.424756 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.453812 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d75d4f7c7-492fj"] Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.457268 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7d75d4f7c7-492fj"] Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.606055 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-679d6c649b-jzn47"] Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.627228 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f9d2899-4e00-467e-afbe-214f40042543" path="/var/lib/kubelet/pods/8f9d2899-4e00-467e-afbe-214f40042543/volumes" Dec 02 01:56:41 crc kubenswrapper[4884]: I1202 01:56:41.627690 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea9acb40-6e2e-4873-9a88-66fb8830d688" path="/var/lib/kubelet/pods/ea9acb40-6e2e-4873-9a88-66fb8830d688/volumes" Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.074602 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.204169 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-5kq68" podUID="ff7b03af-7163-47fa-9397-ce25148d27c5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.119:5353: i/o timeout" Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.367523 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"093e52d3-1df3-49de-97a6-bd8f08bf7a8f","Type":"ContainerStarted","Data":"365c3fe6e851399da4164648cffe6fd37fbdc9a8927b1644bf300c0856806c26"} Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.367577 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"093e52d3-1df3-49de-97a6-bd8f08bf7a8f","Type":"ContainerStarted","Data":"ac4d6c273d54bed796cd6768d8b63a82b838e4a3bb38b86d96f31bdb5233d409"} Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.367588 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"093e52d3-1df3-49de-97a6-bd8f08bf7a8f","Type":"ContainerStarted","Data":"5d62346cada22f97af6bc99b8d3c1377739ca065c83cea2cf5bc6ee3cc09906f"} Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.368179 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.369711 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"2417b2c6-1407-4e68-aed3-b4d14802bc9e","Type":"ContainerStarted","Data":"0abe34c3f054d82b6f308f9fd37d9d645be0e0251e0dcf4d517b750683040533"} Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.371297 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8687f948dd-5zg8m" event={"ID":"19afec6c-37bd-4e1a-9f96-c7e37c1e7748","Type":"ContainerStarted","Data":"dc1a67428c1035adf1127dc8ee764b9cc231e8703a55873d5afebfb59b472e67"} Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.372761 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55fc6c9b58-9v749" event={"ID":"59b51ab4-e0c8-42af-92aa-063e66157aa7","Type":"ContainerStarted","Data":"d1f74407d395e0019fa6a1d81737da09d7d7e55fb3d882e664487b5cbff7aa32"} Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.375468 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wbxwl" event={"ID":"62db24cb-3633-425d-a4b1-f24b4850ce26","Type":"ContainerStarted","Data":"5a9ade1ccf035898476361d1ac54f2318b834323f15eb54b55c2b391543fcb88"} Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.376955 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"25acc2a6-f07c-44c1-aba6-b5f2e38d39ce","Type":"ContainerStarted","Data":"b3112e80a53c32bde79d2db4fc79e22ebb27abe379a66985f6e4d6d75b65ea1f"} Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.382588 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bd7db777-3718-4db3-b144-1fe763e09dae","Type":"ContainerStarted","Data":"38ef9e4270e4246456447362eed8da192d3b0c398ad2072c4435e6d770208a73"} Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.389538 4884 generic.go:334] "Generic (PLEG): container finished" podID="0c25ce5d-09d7-4a8b-be58-baa2d27191e8" containerID="d1297a53a478a98a781b4e0b8c4cb0ad1f0352a7e94c0e82fb854590cfc3ac66" exitCode=0 Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.389615 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-whpnl" event={"ID":"0c25ce5d-09d7-4a8b-be58-baa2d27191e8","Type":"ContainerDied","Data":"d1297a53a478a98a781b4e0b8c4cb0ad1f0352a7e94c0e82fb854590cfc3ac66"} Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.395261 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=2.395242702 podStartE2EDuration="2.395242702s" podCreationTimestamp="2025-12-02 01:56:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:42.3891184 +0000 UTC m=+1219.064955274" watchObservedRunningTime="2025-12-02 01:56:42.395242702 +0000 UTC m=+1219.071079586" Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.400216 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-679d6c649b-jzn47" event={"ID":"321c69df-0c85-4378-af0f-2b127be9b6a9","Type":"ContainerStarted","Data":"c3ab0e53139b21f9409514545ffa66d0e8b941261ba7a639cccb4de8b20b6d66"} Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.400262 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-679d6c649b-jzn47" event={"ID":"321c69df-0c85-4378-af0f-2b127be9b6a9","Type":"ContainerStarted","Data":"1d2295a94cc5d95977813bafce53b7133a267323e5b44d14a37fe3bcdfa4eb4e"} Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.406413 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-wbxwl" podStartSLOduration=25.406399007 podStartE2EDuration="25.406399007s" podCreationTimestamp="2025-12-02 01:56:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:42.404925911 +0000 UTC m=+1219.080762795" watchObservedRunningTime="2025-12-02 01:56:42.406399007 +0000 UTC m=+1219.082235891" Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.897723 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7b6477d857-jzgzr"] Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.899475 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.902794 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.902951 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.912764 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7b6477d857-jzgzr"] Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.997862 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr88h\" (UniqueName: \"kubernetes.io/projected/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-kube-api-access-rr88h\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.997917 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-httpd-config\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.997945 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-combined-ca-bundle\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.997985 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-internal-tls-certs\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.998041 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-ovndb-tls-certs\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.998116 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-config\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:42 crc kubenswrapper[4884]: I1202 01:56:42.998181 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-public-tls-certs\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:43 crc kubenswrapper[4884]: I1202 01:56:43.101359 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr88h\" (UniqueName: \"kubernetes.io/projected/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-kube-api-access-rr88h\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:43 crc kubenswrapper[4884]: I1202 01:56:43.101422 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-httpd-config\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:43 crc kubenswrapper[4884]: I1202 01:56:43.101448 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-combined-ca-bundle\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:43 crc kubenswrapper[4884]: I1202 01:56:43.101504 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-internal-tls-certs\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:43 crc kubenswrapper[4884]: I1202 01:56:43.101588 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-ovndb-tls-certs\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:43 crc kubenswrapper[4884]: I1202 01:56:43.101703 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-config\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:43 crc kubenswrapper[4884]: I1202 01:56:43.101811 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-public-tls-certs\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:43 crc kubenswrapper[4884]: I1202 01:56:43.112407 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-internal-tls-certs\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:43 crc kubenswrapper[4884]: I1202 01:56:43.112644 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-combined-ca-bundle\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:43 crc kubenswrapper[4884]: I1202 01:56:43.114365 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-config\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:43 crc kubenswrapper[4884]: I1202 01:56:43.118818 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-ovndb-tls-certs\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:43 crc kubenswrapper[4884]: I1202 01:56:43.119785 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-httpd-config\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:43 crc kubenswrapper[4884]: I1202 01:56:43.121992 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-public-tls-certs\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:43 crc kubenswrapper[4884]: I1202 01:56:43.124516 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr88h\" (UniqueName: \"kubernetes.io/projected/18e31af0-c0a9-4dc4-8281-5eef1a4477e4-kube-api-access-rr88h\") pod \"neutron-7b6477d857-jzgzr\" (UID: \"18e31af0-c0a9-4dc4-8281-5eef1a4477e4\") " pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:43 crc kubenswrapper[4884]: E1202 01:56:43.156147 4884 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37412014_287b_42f1_9437_ad164aa34f52.slice/crio-decead557aa4909bdaee4b0c705b59f7c4c2cfb3584978e712a87366ff7ab8ac.scope\": RecentStats: unable to find data in memory cache]" Dec 02 01:56:43 crc kubenswrapper[4884]: I1202 01:56:43.235041 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:44 crc kubenswrapper[4884]: I1202 01:56:44.424561 4884 generic.go:334] "Generic (PLEG): container finished" podID="37412014-287b-42f1-9437-ad164aa34f52" containerID="decead557aa4909bdaee4b0c705b59f7c4c2cfb3584978e712a87366ff7ab8ac" exitCode=0 Dec 02 01:56:44 crc kubenswrapper[4884]: I1202 01:56:44.425872 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5vj4w" event={"ID":"37412014-287b-42f1-9437-ad164aa34f52","Type":"ContainerDied","Data":"decead557aa4909bdaee4b0c705b59f7c4c2cfb3584978e712a87366ff7ab8ac"} Dec 02 01:56:44 crc kubenswrapper[4884]: I1202 01:56:44.425941 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f0a9fed-85dd-4c34-8365-a61044cdbf20","Type":"ContainerStarted","Data":"fcfe4db57d6a0c1d480b4cfecb9d1dfcfb8c237a393c0e71d0620d5329c16018"} Dec 02 01:56:44 crc kubenswrapper[4884]: I1202 01:56:44.885467 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.078933 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7b6477d857-jzgzr"] Dec 02 01:56:45 crc kubenswrapper[4884]: W1202 01:56:45.089885 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18e31af0_c0a9_4dc4_8281_5eef1a4477e4.slice/crio-58ed859aeffbc014f1e4ed0e528c637a5d9c31b5be717ec43f5bef6aad27f586 WatchSource:0}: Error finding container 58ed859aeffbc014f1e4ed0e528c637a5d9c31b5be717ec43f5bef6aad27f586: Status 404 returned error can't find the container with id 58ed859aeffbc014f1e4ed0e528c637a5d9c31b5be717ec43f5bef6aad27f586 Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.435607 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-679d6c649b-jzn47" event={"ID":"321c69df-0c85-4378-af0f-2b127be9b6a9","Type":"ContainerStarted","Data":"cedf292b3a4697c44d487a2d956d2b7d11751fa4737d94c1814906ca7baaa879"} Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.436271 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.437346 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8687f948dd-5zg8m" event={"ID":"19afec6c-37bd-4e1a-9f96-c7e37c1e7748","Type":"ContainerStarted","Data":"e9ee6805573eb7480f82a2789ee638b068758db015dd7de5537fc0fd5aa74e61"} Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.439782 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55fc6c9b58-9v749" event={"ID":"59b51ab4-e0c8-42af-92aa-063e66157aa7","Type":"ContainerStarted","Data":"b52f5ae968a87880245e34143fbe5457d552755e5257d2884c65812f7904191b"} Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.442308 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45","Type":"ContainerStarted","Data":"94a6e057cb4483d3a3bf690284e2619434c9de4535aa0e4991c8c869f0dd057e"} Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.443846 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f0a9fed-85dd-4c34-8365-a61044cdbf20","Type":"ContainerStarted","Data":"7132f475a081a80b83ced5da7c6d2d1ae3d04804f96fde26030d855e5fa5423e"} Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.445618 4884 generic.go:334] "Generic (PLEG): container finished" podID="62db24cb-3633-425d-a4b1-f24b4850ce26" containerID="5a9ade1ccf035898476361d1ac54f2318b834323f15eb54b55c2b391543fcb88" exitCode=0 Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.445693 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wbxwl" event={"ID":"62db24cb-3633-425d-a4b1-f24b4850ce26","Type":"ContainerDied","Data":"5a9ade1ccf035898476361d1ac54f2318b834323f15eb54b55c2b391543fcb88"} Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.447142 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-applier-0" event={"ID":"25acc2a6-f07c-44c1-aba6-b5f2e38d39ce","Type":"ContainerStarted","Data":"2dc51e519f3fe4adc64189a0ec119bc864943bea1040f4e1cc6f84bcedc28501"} Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.448975 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bd7db777-3718-4db3-b144-1fe763e09dae","Type":"ContainerStarted","Data":"e49b50a121940b0df2eb8d5518eff658f842152161e4d0a3696201bfa9b1a8fe"} Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.449077 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bd7db777-3718-4db3-b144-1fe763e09dae" containerName="glance-log" containerID="cri-o://38ef9e4270e4246456447362eed8da192d3b0c398ad2072c4435e6d770208a73" gracePeriod=30 Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.449100 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="bd7db777-3718-4db3-b144-1fe763e09dae" containerName="glance-httpd" containerID="cri-o://e49b50a121940b0df2eb8d5518eff658f842152161e4d0a3696201bfa9b1a8fe" gracePeriod=30 Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.453352 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-679d6c649b-jzn47" podStartSLOduration=5.453335922 podStartE2EDuration="5.453335922s" podCreationTimestamp="2025-12-02 01:56:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:45.452547473 +0000 UTC m=+1222.128384377" watchObservedRunningTime="2025-12-02 01:56:45.453335922 +0000 UTC m=+1222.129172826" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.455775 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-whpnl" event={"ID":"0c25ce5d-09d7-4a8b-be58-baa2d27191e8","Type":"ContainerStarted","Data":"9e023e5e5819281374724729f6fbf33e9564e4e8e61e4b446f963b387a9c54d8"} Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.457244 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b6477d857-jzgzr" event={"ID":"18e31af0-c0a9-4dc4-8281-5eef1a4477e4","Type":"ContainerStarted","Data":"58ed859aeffbc014f1e4ed0e528c637a5d9c31b5be717ec43f5bef6aad27f586"} Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.459570 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"2417b2c6-1407-4e68-aed3-b4d14802bc9e","Type":"ContainerStarted","Data":"fb94676af85e01bb962d107a2d9516e111f83b880482244f79a8c359fcf6ee8a"} Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.469133 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.496894 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-8687f948dd-5zg8m" podStartSLOduration=26.999378575 podStartE2EDuration="27.496867827s" podCreationTimestamp="2025-12-02 01:56:18 +0000 UTC" firstStartedPulling="2025-12-02 01:56:41.087344781 +0000 UTC m=+1217.763181665" lastFinishedPulling="2025-12-02 01:56:41.584834033 +0000 UTC m=+1218.260670917" observedRunningTime="2025-12-02 01:56:45.484109112 +0000 UTC m=+1222.159945996" watchObservedRunningTime="2025-12-02 01:56:45.496867827 +0000 UTC m=+1222.172704801" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.551786 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=29.551770153 podStartE2EDuration="29.551770153s" podCreationTimestamp="2025-12-02 01:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:45.549539898 +0000 UTC m=+1222.225376782" watchObservedRunningTime="2025-12-02 01:56:45.551770153 +0000 UTC m=+1222.227607037" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.563663 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-55fc6c9b58-9v749" podStartSLOduration=26.963348535 podStartE2EDuration="27.563649436s" podCreationTimestamp="2025-12-02 01:56:18 +0000 UTC" firstStartedPulling="2025-12-02 01:56:41.067600233 +0000 UTC m=+1217.743437117" lastFinishedPulling="2025-12-02 01:56:41.667901134 +0000 UTC m=+1218.343738018" observedRunningTime="2025-12-02 01:56:45.521114396 +0000 UTC m=+1222.196951280" watchObservedRunningTime="2025-12-02 01:56:45.563649436 +0000 UTC m=+1222.239486320" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.628120 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-applier-0" podStartSLOduration=2.349173184 podStartE2EDuration="5.628093397s" podCreationTimestamp="2025-12-02 01:56:40 +0000 UTC" firstStartedPulling="2025-12-02 01:56:41.438201983 +0000 UTC m=+1218.114038867" lastFinishedPulling="2025-12-02 01:56:44.717122196 +0000 UTC m=+1221.392959080" observedRunningTime="2025-12-02 01:56:45.622457078 +0000 UTC m=+1222.298293972" watchObservedRunningTime="2025-12-02 01:56:45.628093397 +0000 UTC m=+1222.303930311" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.648477 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=2.381518553 podStartE2EDuration="5.64846064s" podCreationTimestamp="2025-12-02 01:56:40 +0000 UTC" firstStartedPulling="2025-12-02 01:56:41.349196806 +0000 UTC m=+1218.025033690" lastFinishedPulling="2025-12-02 01:56:44.616138893 +0000 UTC m=+1221.291975777" observedRunningTime="2025-12-02 01:56:45.643456136 +0000 UTC m=+1222.319293020" watchObservedRunningTime="2025-12-02 01:56:45.64846064 +0000 UTC m=+1222.324297524" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.679090 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55f844cf75-whpnl" podStartSLOduration=5.679073526 podStartE2EDuration="5.679073526s" podCreationTimestamp="2025-12-02 01:56:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:45.676675666 +0000 UTC m=+1222.352512560" watchObservedRunningTime="2025-12-02 01:56:45.679073526 +0000 UTC m=+1222.354910410" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.760661 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.761676 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-applier-0" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.895411 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.971778 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37412014-287b-42f1-9437-ad164aa34f52-logs\") pod \"37412014-287b-42f1-9437-ad164aa34f52\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.971927 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-config-data\") pod \"37412014-287b-42f1-9437-ad164aa34f52\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.971960 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbh55\" (UniqueName: \"kubernetes.io/projected/37412014-287b-42f1-9437-ad164aa34f52-kube-api-access-kbh55\") pod \"37412014-287b-42f1-9437-ad164aa34f52\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.971995 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-combined-ca-bundle\") pod \"37412014-287b-42f1-9437-ad164aa34f52\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.972126 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-scripts\") pod \"37412014-287b-42f1-9437-ad164aa34f52\" (UID: \"37412014-287b-42f1-9437-ad164aa34f52\") " Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.972318 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37412014-287b-42f1-9437-ad164aa34f52-logs" (OuterVolumeSpecName: "logs") pod "37412014-287b-42f1-9437-ad164aa34f52" (UID: "37412014-287b-42f1-9437-ad164aa34f52"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.972631 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37412014-287b-42f1-9437-ad164aa34f52-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.976600 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37412014-287b-42f1-9437-ad164aa34f52-kube-api-access-kbh55" (OuterVolumeSpecName: "kube-api-access-kbh55") pod "37412014-287b-42f1-9437-ad164aa34f52" (UID: "37412014-287b-42f1-9437-ad164aa34f52"). InnerVolumeSpecName "kube-api-access-kbh55". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:45 crc kubenswrapper[4884]: I1202 01:56:45.979146 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-scripts" (OuterVolumeSpecName: "scripts") pod "37412014-287b-42f1-9437-ad164aa34f52" (UID: "37412014-287b-42f1-9437-ad164aa34f52"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.000601 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37412014-287b-42f1-9437-ad164aa34f52" (UID: "37412014-287b-42f1-9437-ad164aa34f52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.002525 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-config-data" (OuterVolumeSpecName: "config-data") pod "37412014-287b-42f1-9437-ad164aa34f52" (UID: "37412014-287b-42f1-9437-ad164aa34f52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.074828 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.074887 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbh55\" (UniqueName: \"kubernetes.io/projected/37412014-287b-42f1-9437-ad164aa34f52-kube-api-access-kbh55\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.074905 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.074917 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37412014-287b-42f1-9437-ad164aa34f52-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.490629 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-5vj4w" event={"ID":"37412014-287b-42f1-9437-ad164aa34f52","Type":"ContainerDied","Data":"b3f53e3e6bc800ed5e6e0b38aa6149910b413086b3aab804c20fc459fc543c04"} Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.491005 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3f53e3e6bc800ed5e6e0b38aa6149910b413086b3aab804c20fc459fc543c04" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.490901 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-5vj4w" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.515902 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b6477d857-jzgzr" event={"ID":"18e31af0-c0a9-4dc4-8281-5eef1a4477e4","Type":"ContainerStarted","Data":"50482264b0a724af41449fae2ac9f0d5d0043f118d76e418f5894081c026ba48"} Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.515947 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7b6477d857-jzgzr" event={"ID":"18e31af0-c0a9-4dc4-8281-5eef1a4477e4","Type":"ContainerStarted","Data":"8a84386803131c43f3482c214c28340dea24073107e2f9b89f0004b3172c092c"} Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.516836 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.554333 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7b6477d857-jzgzr" podStartSLOduration=4.554316075 podStartE2EDuration="4.554316075s" podCreationTimestamp="2025-12-02 01:56:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:46.54927728 +0000 UTC m=+1223.225114164" watchObservedRunningTime="2025-12-02 01:56:46.554316075 +0000 UTC m=+1223.230152959" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.557054 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f0a9fed-85dd-4c34-8365-a61044cdbf20","Type":"ContainerStarted","Data":"fd7879461e7dca7abb24ca95ef28825ad5e2efb3384758e92bdb5f265951a47b"} Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.557257 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0f0a9fed-85dd-4c34-8365-a61044cdbf20" containerName="glance-log" containerID="cri-o://7132f475a081a80b83ced5da7c6d2d1ae3d04804f96fde26030d855e5fa5423e" gracePeriod=30 Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.557822 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0f0a9fed-85dd-4c34-8365-a61044cdbf20" containerName="glance-httpd" containerID="cri-o://fd7879461e7dca7abb24ca95ef28825ad5e2efb3384758e92bdb5f265951a47b" gracePeriod=30 Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.605024 4884 generic.go:334] "Generic (PLEG): container finished" podID="bd7db777-3718-4db3-b144-1fe763e09dae" containerID="e49b50a121940b0df2eb8d5518eff658f842152161e4d0a3696201bfa9b1a8fe" exitCode=143 Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.605072 4884 generic.go:334] "Generic (PLEG): container finished" podID="bd7db777-3718-4db3-b144-1fe763e09dae" containerID="38ef9e4270e4246456447362eed8da192d3b0c398ad2072c4435e6d770208a73" exitCode=143 Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.605296 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bd7db777-3718-4db3-b144-1fe763e09dae","Type":"ContainerDied","Data":"e49b50a121940b0df2eb8d5518eff658f842152161e4d0a3696201bfa9b1a8fe"} Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.605326 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bd7db777-3718-4db3-b144-1fe763e09dae","Type":"ContainerDied","Data":"38ef9e4270e4246456447362eed8da192d3b0c398ad2072c4435e6d770208a73"} Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.652520 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-595d8586d8-j5926"] Dec 02 01:56:46 crc kubenswrapper[4884]: E1202 01:56:46.653381 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37412014-287b-42f1-9437-ad164aa34f52" containerName="placement-db-sync" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.653399 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="37412014-287b-42f1-9437-ad164aa34f52" containerName="placement-db-sync" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.653760 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="37412014-287b-42f1-9437-ad164aa34f52" containerName="placement-db-sync" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.655352 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.678567 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=30.678543632 podStartE2EDuration="30.678543632s" podCreationTimestamp="2025-12-02 01:56:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:46.594768923 +0000 UTC m=+1223.270605817" watchObservedRunningTime="2025-12-02 01:56:46.678543632 +0000 UTC m=+1223.354380516" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.686451 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.686727 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-q4l62" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.686913 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.687142 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.688000 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.729211 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f4a5e6c-067d-4f50-b053-950fb802d52e-internal-tls-certs\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.729496 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f4a5e6c-067d-4f50-b053-950fb802d52e-config-data\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.729551 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f4a5e6c-067d-4f50-b053-950fb802d52e-scripts\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.729571 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f4a5e6c-067d-4f50-b053-950fb802d52e-logs\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.729663 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f4a5e6c-067d-4f50-b053-950fb802d52e-combined-ca-bundle\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.729711 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5vf5\" (UniqueName: \"kubernetes.io/projected/1f4a5e6c-067d-4f50-b053-950fb802d52e-kube-api-access-w5vf5\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.730962 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f4a5e6c-067d-4f50-b053-950fb802d52e-public-tls-certs\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.743456 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-595d8586d8-j5926"] Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.842921 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f4a5e6c-067d-4f50-b053-950fb802d52e-config-data\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.843153 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f4a5e6c-067d-4f50-b053-950fb802d52e-scripts\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.843170 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f4a5e6c-067d-4f50-b053-950fb802d52e-logs\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.843202 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f4a5e6c-067d-4f50-b053-950fb802d52e-combined-ca-bundle\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.843221 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5vf5\" (UniqueName: \"kubernetes.io/projected/1f4a5e6c-067d-4f50-b053-950fb802d52e-kube-api-access-w5vf5\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.843288 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f4a5e6c-067d-4f50-b053-950fb802d52e-public-tls-certs\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.843323 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f4a5e6c-067d-4f50-b053-950fb802d52e-internal-tls-certs\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.844849 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1f4a5e6c-067d-4f50-b053-950fb802d52e-logs\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.857340 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f4a5e6c-067d-4f50-b053-950fb802d52e-public-tls-certs\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.857650 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f4a5e6c-067d-4f50-b053-950fb802d52e-scripts\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.859988 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f4a5e6c-067d-4f50-b053-950fb802d52e-config-data\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.860445 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f4a5e6c-067d-4f50-b053-950fb802d52e-internal-tls-certs\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.868725 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f4a5e6c-067d-4f50-b053-950fb802d52e-combined-ca-bundle\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.871142 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5vf5\" (UniqueName: \"kubernetes.io/projected/1f4a5e6c-067d-4f50-b053-950fb802d52e-kube-api-access-w5vf5\") pod \"placement-595d8586d8-j5926\" (UID: \"1f4a5e6c-067d-4f50-b053-950fb802d52e\") " pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.934701 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.971714 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.971781 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:56:46 crc kubenswrapper[4884]: I1202 01:56:46.998241 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.046695 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-internal-tls-certs\") pod \"bd7db777-3718-4db3-b144-1fe763e09dae\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.046736 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd7db777-3718-4db3-b144-1fe763e09dae-httpd-run\") pod \"bd7db777-3718-4db3-b144-1fe763e09dae\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.046806 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-combined-ca-bundle\") pod \"bd7db777-3718-4db3-b144-1fe763e09dae\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.047500 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd7db777-3718-4db3-b144-1fe763e09dae-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bd7db777-3718-4db3-b144-1fe763e09dae" (UID: "bd7db777-3718-4db3-b144-1fe763e09dae"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.047560 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd7db777-3718-4db3-b144-1fe763e09dae-logs\") pod \"bd7db777-3718-4db3-b144-1fe763e09dae\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.047592 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"bd7db777-3718-4db3-b144-1fe763e09dae\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.047843 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd7db777-3718-4db3-b144-1fe763e09dae-logs" (OuterVolumeSpecName: "logs") pod "bd7db777-3718-4db3-b144-1fe763e09dae" (UID: "bd7db777-3718-4db3-b144-1fe763e09dae"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.047936 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-config-data\") pod \"bd7db777-3718-4db3-b144-1fe763e09dae\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.048317 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-scripts\") pod \"bd7db777-3718-4db3-b144-1fe763e09dae\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.048350 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdhls\" (UniqueName: \"kubernetes.io/projected/bd7db777-3718-4db3-b144-1fe763e09dae-kube-api-access-jdhls\") pod \"bd7db777-3718-4db3-b144-1fe763e09dae\" (UID: \"bd7db777-3718-4db3-b144-1fe763e09dae\") " Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.048705 4884 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bd7db777-3718-4db3-b144-1fe763e09dae-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.048721 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bd7db777-3718-4db3-b144-1fe763e09dae-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.055920 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "bd7db777-3718-4db3-b144-1fe763e09dae" (UID: "bd7db777-3718-4db3-b144-1fe763e09dae"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.055948 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd7db777-3718-4db3-b144-1fe763e09dae-kube-api-access-jdhls" (OuterVolumeSpecName: "kube-api-access-jdhls") pod "bd7db777-3718-4db3-b144-1fe763e09dae" (UID: "bd7db777-3718-4db3-b144-1fe763e09dae"). InnerVolumeSpecName "kube-api-access-jdhls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.055975 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-scripts" (OuterVolumeSpecName: "scripts") pod "bd7db777-3718-4db3-b144-1fe763e09dae" (UID: "bd7db777-3718-4db3-b144-1fe763e09dae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.080372 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd7db777-3718-4db3-b144-1fe763e09dae" (UID: "bd7db777-3718-4db3-b144-1fe763e09dae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.106957 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bd7db777-3718-4db3-b144-1fe763e09dae" (UID: "bd7db777-3718-4db3-b144-1fe763e09dae"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.128140 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-config-data" (OuterVolumeSpecName: "config-data") pod "bd7db777-3718-4db3-b144-1fe763e09dae" (UID: "bd7db777-3718-4db3-b144-1fe763e09dae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.150931 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.150963 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdhls\" (UniqueName: \"kubernetes.io/projected/bd7db777-3718-4db3-b144-1fe763e09dae-kube-api-access-jdhls\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.150975 4884 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.150984 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.151013 4884 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.151021 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd7db777-3718-4db3-b144-1fe763e09dae-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.175162 4884 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.254180 4884 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.348677 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.348952 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.619191 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.625645 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"bd7db777-3718-4db3-b144-1fe763e09dae","Type":"ContainerDied","Data":"aaed6e805751b822c9c9f96bfff8ff9a0ec8cde86e69d3256920878330d61bc0"} Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.625720 4884 scope.go:117] "RemoveContainer" containerID="e49b50a121940b0df2eb8d5518eff658f842152161e4d0a3696201bfa9b1a8fe" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.626207 4884 generic.go:334] "Generic (PLEG): container finished" podID="0f0a9fed-85dd-4c34-8365-a61044cdbf20" containerID="7132f475a081a80b83ced5da7c6d2d1ae3d04804f96fde26030d855e5fa5423e" exitCode=143 Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.632460 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f0a9fed-85dd-4c34-8365-a61044cdbf20","Type":"ContainerDied","Data":"7132f475a081a80b83ced5da7c6d2d1ae3d04804f96fde26030d855e5fa5423e"} Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.668071 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.686461 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.692863 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:56:47 crc kubenswrapper[4884]: E1202 01:56:47.693286 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd7db777-3718-4db3-b144-1fe763e09dae" containerName="glance-log" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.693298 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd7db777-3718-4db3-b144-1fe763e09dae" containerName="glance-log" Dec 02 01:56:47 crc kubenswrapper[4884]: E1202 01:56:47.693314 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd7db777-3718-4db3-b144-1fe763e09dae" containerName="glance-httpd" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.693321 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd7db777-3718-4db3-b144-1fe763e09dae" containerName="glance-httpd" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.693498 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd7db777-3718-4db3-b144-1fe763e09dae" containerName="glance-log" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.693531 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd7db777-3718-4db3-b144-1fe763e09dae" containerName="glance-httpd" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.694548 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.696521 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.696549 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.713468 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.868127 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.868182 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fdad2f7-f276-458e-96dd-6cc7fda79cac-logs\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.868214 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.868283 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.868365 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.868764 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27fl9\" (UniqueName: \"kubernetes.io/projected/5fdad2f7-f276-458e-96dd-6cc7fda79cac-kube-api-access-27fl9\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.868799 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.868824 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5fdad2f7-f276-458e-96dd-6cc7fda79cac-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.970511 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.970555 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27fl9\" (UniqueName: \"kubernetes.io/projected/5fdad2f7-f276-458e-96dd-6cc7fda79cac-kube-api-access-27fl9\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.970580 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5fdad2f7-f276-458e-96dd-6cc7fda79cac-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.970622 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.970652 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fdad2f7-f276-458e-96dd-6cc7fda79cac-logs\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.970678 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.970696 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.970724 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.971078 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.971946 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fdad2f7-f276-458e-96dd-6cc7fda79cac-logs\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.972435 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5fdad2f7-f276-458e-96dd-6cc7fda79cac-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.978473 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.978991 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.979338 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.983513 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:47 crc kubenswrapper[4884]: I1202 01:56:47.992876 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27fl9\" (UniqueName: \"kubernetes.io/projected/5fdad2f7-f276-458e-96dd-6cc7fda79cac-kube-api-access-27fl9\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:48 crc kubenswrapper[4884]: I1202 01:56:48.009188 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:56:48 crc kubenswrapper[4884]: I1202 01:56:48.309325 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:49 crc kubenswrapper[4884]: I1202 01:56:49.135185 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:49 crc kubenswrapper[4884]: I1202 01:56:49.135275 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:56:49 crc kubenswrapper[4884]: I1202 01:56:49.198685 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:49 crc kubenswrapper[4884]: I1202 01:56:49.198840 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:56:49 crc kubenswrapper[4884]: I1202 01:56:49.632979 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd7db777-3718-4db3-b144-1fe763e09dae" path="/var/lib/kubelet/pods/bd7db777-3718-4db3-b144-1fe763e09dae/volumes" Dec 02 01:56:49 crc kubenswrapper[4884]: I1202 01:56:49.649774 4884 generic.go:334] "Generic (PLEG): container finished" podID="0f0a9fed-85dd-4c34-8365-a61044cdbf20" containerID="fd7879461e7dca7abb24ca95ef28825ad5e2efb3384758e92bdb5f265951a47b" exitCode=143 Dec 02 01:56:49 crc kubenswrapper[4884]: I1202 01:56:49.649909 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f0a9fed-85dd-4c34-8365-a61044cdbf20","Type":"ContainerDied","Data":"fd7879461e7dca7abb24ca95ef28825ad5e2efb3384758e92bdb5f265951a47b"} Dec 02 01:56:49 crc kubenswrapper[4884]: I1202 01:56:49.956226 4884 scope.go:117] "RemoveContainer" containerID="38ef9e4270e4246456447362eed8da192d3b0c398ad2072c4435e6d770208a73" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.071151 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.236827 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-credential-keys\") pod \"62db24cb-3633-425d-a4b1-f24b4850ce26\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.237214 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f686p\" (UniqueName: \"kubernetes.io/projected/62db24cb-3633-425d-a4b1-f24b4850ce26-kube-api-access-f686p\") pod \"62db24cb-3633-425d-a4b1-f24b4850ce26\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.237294 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-scripts\") pod \"62db24cb-3633-425d-a4b1-f24b4850ce26\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.237340 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-config-data\") pod \"62db24cb-3633-425d-a4b1-f24b4850ce26\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.237405 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-combined-ca-bundle\") pod \"62db24cb-3633-425d-a4b1-f24b4850ce26\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.237442 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-fernet-keys\") pod \"62db24cb-3633-425d-a4b1-f24b4850ce26\" (UID: \"62db24cb-3633-425d-a4b1-f24b4850ce26\") " Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.243877 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-scripts" (OuterVolumeSpecName: "scripts") pod "62db24cb-3633-425d-a4b1-f24b4850ce26" (UID: "62db24cb-3633-425d-a4b1-f24b4850ce26"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.243919 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "62db24cb-3633-425d-a4b1-f24b4850ce26" (UID: "62db24cb-3633-425d-a4b1-f24b4850ce26"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.255213 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "62db24cb-3633-425d-a4b1-f24b4850ce26" (UID: "62db24cb-3633-425d-a4b1-f24b4850ce26"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.265718 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-config-data" (OuterVolumeSpecName: "config-data") pod "62db24cb-3633-425d-a4b1-f24b4850ce26" (UID: "62db24cb-3633-425d-a4b1-f24b4850ce26"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.278943 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62db24cb-3633-425d-a4b1-f24b4850ce26-kube-api-access-f686p" (OuterVolumeSpecName: "kube-api-access-f686p") pod "62db24cb-3633-425d-a4b1-f24b4850ce26" (UID: "62db24cb-3633-425d-a4b1-f24b4850ce26"). InnerVolumeSpecName "kube-api-access-f686p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.281640 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62db24cb-3633-425d-a4b1-f24b4850ce26" (UID: "62db24cb-3633-425d-a4b1-f24b4850ce26"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.339853 4884 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.339887 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f686p\" (UniqueName: \"kubernetes.io/projected/62db24cb-3633-425d-a4b1-f24b4850ce26-kube-api-access-f686p\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.339897 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.339907 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.339915 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.339925 4884 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/62db24cb-3633-425d-a4b1-f24b4850ce26-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.354103 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.413374 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-crspk"] Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.413631 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" podUID="4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" containerName="dnsmasq-dns" containerID="cri-o://8158cc45cc0365e890619d1e6f8e9ada649c3e8ae14caa6a07bd7153ed7ecd78" gracePeriod=10 Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.552437 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.575054 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-595d8586d8-j5926"] Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.667527 4884 generic.go:334] "Generic (PLEG): container finished" podID="4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" containerID="8158cc45cc0365e890619d1e6f8e9ada649c3e8ae14caa6a07bd7153ed7ecd78" exitCode=0 Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.667789 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" event={"ID":"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd","Type":"ContainerDied","Data":"8158cc45cc0365e890619d1e6f8e9ada649c3e8ae14caa6a07bd7153ed7ecd78"} Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.668905 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-wbxwl" event={"ID":"62db24cb-3633-425d-a4b1-f24b4850ce26","Type":"ContainerDied","Data":"a36efa1a0b410641f38a458cf14451ad053341e02c79d177546d9af853567539"} Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.668926 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a36efa1a0b410641f38a458cf14451ad053341e02c79d177546d9af853567539" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.668984 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-wbxwl" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.691380 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.720822 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.720890 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-applier-0" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.723882 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.728799 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Dec 02 01:56:50 crc kubenswrapper[4884]: I1202 01:56:50.755515 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-applier-0" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.188727 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-87844dc44-b7fcj"] Dec 02 01:56:51 crc kubenswrapper[4884]: E1202 01:56:51.189192 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62db24cb-3633-425d-a4b1-f24b4850ce26" containerName="keystone-bootstrap" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.189213 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="62db24cb-3633-425d-a4b1-f24b4850ce26" containerName="keystone-bootstrap" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.189421 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="62db24cb-3633-425d-a4b1-f24b4850ce26" containerName="keystone-bootstrap" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.190163 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.195976 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.196160 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.196266 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-42pg8" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.196367 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.196514 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.196619 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.221557 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-87844dc44-b7fcj"] Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.277828 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-fernet-keys\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.277968 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-scripts\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.277993 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq76b\" (UniqueName: \"kubernetes.io/projected/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-kube-api-access-jq76b\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.278020 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-combined-ca-bundle\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.278045 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-config-data\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.278071 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-public-tls-certs\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.278174 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-internal-tls-certs\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.278213 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-credential-keys\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.379592 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-public-tls-certs\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.380023 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-internal-tls-certs\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.380063 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-credential-keys\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.380124 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-fernet-keys\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.380199 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq76b\" (UniqueName: \"kubernetes.io/projected/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-kube-api-access-jq76b\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.380219 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-scripts\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.380237 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-combined-ca-bundle\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.380256 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-config-data\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.472990 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-config-data\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.473761 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-scripts\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.479066 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-internal-tls-certs\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.480527 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-combined-ca-bundle\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.485240 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-public-tls-certs\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.486983 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-credential-keys\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.489484 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-fernet-keys\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.489761 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq76b\" (UniqueName: \"kubernetes.io/projected/8fad23d3-2fad-430e-a78f-7cb9e778f9ae-kube-api-access-jq76b\") pod \"keystone-87844dc44-b7fcj\" (UID: \"8fad23d3-2fad-430e-a78f-7cb9e778f9ae\") " pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.527412 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.686098 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.734018 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.821983 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-applier-0" Dec 02 01:56:51 crc kubenswrapper[4884]: I1202 01:56:51.827422 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Dec 02 01:56:53 crc kubenswrapper[4884]: W1202 01:56:53.021658 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f4a5e6c_067d_4f50_b053_950fb802d52e.slice/crio-39e0bb83be8434a6f73464f03ef16af6edf30ec182486a6a7a807edd04efe496 WatchSource:0}: Error finding container 39e0bb83be8434a6f73464f03ef16af6edf30ec182486a6a7a807edd04efe496: Status 404 returned error can't find the container with id 39e0bb83be8434a6f73464f03ef16af6edf30ec182486a6a7a807edd04efe496 Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.279597 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.438394 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-public-tls-certs\") pod \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.438667 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx882\" (UniqueName: \"kubernetes.io/projected/0f0a9fed-85dd-4c34-8365-a61044cdbf20-kube-api-access-gx882\") pod \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.438705 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-combined-ca-bundle\") pod \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.438803 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f0a9fed-85dd-4c34-8365-a61044cdbf20-logs\") pod \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.438819 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.438845 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f0a9fed-85dd-4c34-8365-a61044cdbf20-httpd-run\") pod \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.438864 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-config-data\") pod \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.438888 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-scripts\") pod \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\" (UID: \"0f0a9fed-85dd-4c34-8365-a61044cdbf20\") " Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.440602 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f0a9fed-85dd-4c34-8365-a61044cdbf20-logs" (OuterVolumeSpecName: "logs") pod "0f0a9fed-85dd-4c34-8365-a61044cdbf20" (UID: "0f0a9fed-85dd-4c34-8365-a61044cdbf20"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.441979 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f0a9fed-85dd-4c34-8365-a61044cdbf20-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0f0a9fed-85dd-4c34-8365-a61044cdbf20" (UID: "0f0a9fed-85dd-4c34-8365-a61044cdbf20"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.447287 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "0f0a9fed-85dd-4c34-8365-a61044cdbf20" (UID: "0f0a9fed-85dd-4c34-8365-a61044cdbf20"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.454160 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-scripts" (OuterVolumeSpecName: "scripts") pod "0f0a9fed-85dd-4c34-8365-a61044cdbf20" (UID: "0f0a9fed-85dd-4c34-8365-a61044cdbf20"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.460924 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f0a9fed-85dd-4c34-8365-a61044cdbf20-kube-api-access-gx882" (OuterVolumeSpecName: "kube-api-access-gx882") pod "0f0a9fed-85dd-4c34-8365-a61044cdbf20" (UID: "0f0a9fed-85dd-4c34-8365-a61044cdbf20"). InnerVolumeSpecName "kube-api-access-gx882". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.508814 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f0a9fed-85dd-4c34-8365-a61044cdbf20" (UID: "0f0a9fed-85dd-4c34-8365-a61044cdbf20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.510186 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.542402 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx882\" (UniqueName: \"kubernetes.io/projected/0f0a9fed-85dd-4c34-8365-a61044cdbf20-kube-api-access-gx882\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.542435 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.542446 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f0a9fed-85dd-4c34-8365-a61044cdbf20-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.542468 4884 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.542476 4884 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0f0a9fed-85dd-4c34-8365-a61044cdbf20-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.542486 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.556834 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0f0a9fed-85dd-4c34-8365-a61044cdbf20" (UID: "0f0a9fed-85dd-4c34-8365-a61044cdbf20"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.583543 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-config-data" (OuterVolumeSpecName: "config-data") pod "0f0a9fed-85dd-4c34-8365-a61044cdbf20" (UID: "0f0a9fed-85dd-4c34-8365-a61044cdbf20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.585242 4884 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.650567 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-config\") pod \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.650867 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-dns-swift-storage-0\") pod \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.650958 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-ovsdbserver-sb\") pod \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.651024 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fn5hp\" (UniqueName: \"kubernetes.io/projected/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-kube-api-access-fn5hp\") pod \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.651062 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-dns-svc\") pod \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.651136 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-ovsdbserver-nb\") pod \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\" (UID: \"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd\") " Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.651486 4884 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.651498 4884 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.651507 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f0a9fed-85dd-4c34-8365-a61044cdbf20-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.657764 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-kube-api-access-fn5hp" (OuterVolumeSpecName: "kube-api-access-fn5hp") pod "4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" (UID: "4f6fe6c3-c15e-45d3-91e5-1474030f2bbd"). InnerVolumeSpecName "kube-api-access-fn5hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.691955 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-87844dc44-b7fcj"] Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.710934 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5fdad2f7-f276-458e-96dd-6cc7fda79cac","Type":"ContainerStarted","Data":"6ce0a59277c766ef0cb22193f407931d078d6a831472bf42e83ab6d2d475f76e"} Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.724827 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" (UID: "4f6fe6c3-c15e-45d3-91e5-1474030f2bbd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.726240 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wmhnq" event={"ID":"fe8c221e-e320-4358-b940-436f38a423d3","Type":"ContainerStarted","Data":"1989683bf093eb5ecb39497dbe41e94fa661c162853fe0b1a5047b3edb0cbeb1"} Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.729327 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" event={"ID":"4f6fe6c3-c15e-45d3-91e5-1474030f2bbd","Type":"ContainerDied","Data":"038dd18c1e37350d8c154f41719ca2df601d7ef0bcd8c1e2f405f1491caa4127"} Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.729460 4884 scope.go:117] "RemoveContainer" containerID="8158cc45cc0365e890619d1e6f8e9ada649c3e8ae14caa6a07bd7153ed7ecd78" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.729577 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.733661 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-595d8586d8-j5926" event={"ID":"1f4a5e6c-067d-4f50-b053-950fb802d52e","Type":"ContainerStarted","Data":"301ff72af72b42b7cac7c0d804b47e0a949313a2682d97575650864b4bd766cf"} Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.733706 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-595d8586d8-j5926" event={"ID":"1f4a5e6c-067d-4f50-b053-950fb802d52e","Type":"ContainerStarted","Data":"39e0bb83be8434a6f73464f03ef16af6edf30ec182486a6a7a807edd04efe496"} Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.736776 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" (UID: "4f6fe6c3-c15e-45d3-91e5-1474030f2bbd"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.740423 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.740503 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0f0a9fed-85dd-4c34-8365-a61044cdbf20","Type":"ContainerDied","Data":"fcfe4db57d6a0c1d480b4cfecb9d1dfcfb8c237a393c0e71d0620d5329c16018"} Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.753354 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fn5hp\" (UniqueName: \"kubernetes.io/projected/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-kube-api-access-fn5hp\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.753386 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.753395 4884 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.756247 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-wmhnq" podStartSLOduration=3.340496762 podStartE2EDuration="43.756230183s" podCreationTimestamp="2025-12-02 01:56:10 +0000 UTC" firstStartedPulling="2025-12-02 01:56:12.806912778 +0000 UTC m=+1189.482749662" lastFinishedPulling="2025-12-02 01:56:53.222646199 +0000 UTC m=+1229.898483083" observedRunningTime="2025-12-02 01:56:53.7447802 +0000 UTC m=+1230.420617084" watchObservedRunningTime="2025-12-02 01:56:53.756230183 +0000 UTC m=+1230.432067067" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.760092 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-config" (OuterVolumeSpecName: "config") pod "4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" (UID: "4f6fe6c3-c15e-45d3-91e5-1474030f2bbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.770354 4884 scope.go:117] "RemoveContainer" containerID="544956c368c052542d947e912d089b6be064eddfad9ce9e6c447e3200a3cf0e8" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.783289 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" (UID: "4f6fe6c3-c15e-45d3-91e5-1474030f2bbd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.784645 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.805858 4884 scope.go:117] "RemoveContainer" containerID="fd7879461e7dca7abb24ca95ef28825ad5e2efb3384758e92bdb5f265951a47b" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.814446 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.815929 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" (UID: "4f6fe6c3-c15e-45d3-91e5-1474030f2bbd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.848058 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:56:53 crc kubenswrapper[4884]: E1202 01:56:53.848439 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f0a9fed-85dd-4c34-8365-a61044cdbf20" containerName="glance-httpd" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.848450 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f0a9fed-85dd-4c34-8365-a61044cdbf20" containerName="glance-httpd" Dec 02 01:56:53 crc kubenswrapper[4884]: E1202 01:56:53.848482 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f0a9fed-85dd-4c34-8365-a61044cdbf20" containerName="glance-log" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.848489 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f0a9fed-85dd-4c34-8365-a61044cdbf20" containerName="glance-log" Dec 02 01:56:53 crc kubenswrapper[4884]: E1202 01:56:53.848506 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" containerName="init" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.848511 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" containerName="init" Dec 02 01:56:53 crc kubenswrapper[4884]: E1202 01:56:53.848525 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" containerName="dnsmasq-dns" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.848532 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" containerName="dnsmasq-dns" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.848696 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" containerName="dnsmasq-dns" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.848711 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f0a9fed-85dd-4c34-8365-a61044cdbf20" containerName="glance-log" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.848722 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f0a9fed-85dd-4c34-8365-a61044cdbf20" containerName="glance-httpd" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.849774 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.852850 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.852862 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.854640 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.854662 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.854672 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.890828 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.893939 4884 scope.go:117] "RemoveContainer" containerID="7132f475a081a80b83ced5da7c6d2d1ae3d04804f96fde26030d855e5fa5423e" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.960705 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5702eb-cee9-4049-b3b4-601f8f5e1854-logs\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.960831 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.960894 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx5mr\" (UniqueName: \"kubernetes.io/projected/fb5702eb-cee9-4049-b3b4-601f8f5e1854-kube-api-access-zx5mr\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.960925 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-scripts\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.960948 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb5702eb-cee9-4049-b3b4-601f8f5e1854-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.960970 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.960998 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-config-data\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:53 crc kubenswrapper[4884]: I1202 01:56:53.961012 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.062651 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.062975 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5702eb-cee9-4049-b3b4-601f8f5e1854-logs\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.063025 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.063083 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx5mr\" (UniqueName: \"kubernetes.io/projected/fb5702eb-cee9-4049-b3b4-601f8f5e1854-kube-api-access-zx5mr\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.063108 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-scripts\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.063130 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb5702eb-cee9-4049-b3b4-601f8f5e1854-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.063152 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.063185 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-config-data\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.073228 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5702eb-cee9-4049-b3b4-601f8f5e1854-logs\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.073595 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.076783 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb5702eb-cee9-4049-b3b4-601f8f5e1854-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.086327 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-scripts\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.087801 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.088043 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-config-data\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.088369 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx5mr\" (UniqueName: \"kubernetes.io/projected/fb5702eb-cee9-4049-b3b4-601f8f5e1854-kube-api-access-zx5mr\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.100341 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.124559 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.187452 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.260654 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-crspk"] Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.264577 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-crspk"] Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.760496 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-595d8586d8-j5926" event={"ID":"1f4a5e6c-067d-4f50-b053-950fb802d52e","Type":"ContainerStarted","Data":"e205b6bd9118c13e9d4c2d3f46dfc2c3ac72e2a57ca022cfccaaf6b8c4101434"} Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.762179 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.762213 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-595d8586d8-j5926" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.765040 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-87844dc44-b7fcj" event={"ID":"8fad23d3-2fad-430e-a78f-7cb9e778f9ae","Type":"ContainerStarted","Data":"b2d7132b7a135fc3ab0730f7844f14a67d3de4c83af05eff5b5426d88d401974"} Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.765079 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-87844dc44-b7fcj" event={"ID":"8fad23d3-2fad-430e-a78f-7cb9e778f9ae","Type":"ContainerStarted","Data":"a1555c651486387936c7f30f956b0ff3f6f780ff2548a9e65562ebc8bf994dff"} Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.765211 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.768191 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5fdad2f7-f276-458e-96dd-6cc7fda79cac","Type":"ContainerStarted","Data":"abe05486c9de991f37c1e7d97c6cddbef0c9116bc13387312390c4b258d677fb"} Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.772374 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:56:54 crc kubenswrapper[4884]: W1202 01:56:54.788873 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb5702eb_cee9_4049_b3b4_601f8f5e1854.slice/crio-06a8b59e950e9d35a57f05254d11958b7f2b55b279c3bdc1ddcdbbebfe783b2c WatchSource:0}: Error finding container 06a8b59e950e9d35a57f05254d11958b7f2b55b279c3bdc1ddcdbbebfe783b2c: Status 404 returned error can't find the container with id 06a8b59e950e9d35a57f05254d11958b7f2b55b279c3bdc1ddcdbbebfe783b2c Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.793810 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-595d8586d8-j5926" podStartSLOduration=8.793791818999999 podStartE2EDuration="8.793791819s" podCreationTimestamp="2025-12-02 01:56:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:54.783556547 +0000 UTC m=+1231.459393451" watchObservedRunningTime="2025-12-02 01:56:54.793791819 +0000 UTC m=+1231.469628703" Dec 02 01:56:54 crc kubenswrapper[4884]: I1202 01:56:54.810677 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-87844dc44-b7fcj" podStartSLOduration=3.810655636 podStartE2EDuration="3.810655636s" podCreationTimestamp="2025-12-02 01:56:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:54.801766366 +0000 UTC m=+1231.477603260" watchObservedRunningTime="2025-12-02 01:56:54.810655636 +0000 UTC m=+1231.486492520" Dec 02 01:56:55 crc kubenswrapper[4884]: I1202 01:56:55.014428 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Dec 02 01:56:55 crc kubenswrapper[4884]: I1202 01:56:55.014984 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="093e52d3-1df3-49de-97a6-bd8f08bf7a8f" containerName="watcher-api-log" containerID="cri-o://ac4d6c273d54bed796cd6768d8b63a82b838e4a3bb38b86d96f31bdb5233d409" gracePeriod=30 Dec 02 01:56:55 crc kubenswrapper[4884]: I1202 01:56:55.015412 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-api-0" podUID="093e52d3-1df3-49de-97a6-bd8f08bf7a8f" containerName="watcher-api" containerID="cri-o://365c3fe6e851399da4164648cffe6fd37fbdc9a8927b1644bf300c0856806c26" gracePeriod=30 Dec 02 01:56:55 crc kubenswrapper[4884]: I1202 01:56:55.639390 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f0a9fed-85dd-4c34-8365-a61044cdbf20" path="/var/lib/kubelet/pods/0f0a9fed-85dd-4c34-8365-a61044cdbf20/volumes" Dec 02 01:56:55 crc kubenswrapper[4884]: I1202 01:56:55.642461 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" path="/var/lib/kubelet/pods/4f6fe6c3-c15e-45d3-91e5-1474030f2bbd/volumes" Dec 02 01:56:55 crc kubenswrapper[4884]: I1202 01:56:55.780297 4884 generic.go:334] "Generic (PLEG): container finished" podID="093e52d3-1df3-49de-97a6-bd8f08bf7a8f" containerID="ac4d6c273d54bed796cd6768d8b63a82b838e4a3bb38b86d96f31bdb5233d409" exitCode=143 Dec 02 01:56:55 crc kubenswrapper[4884]: I1202 01:56:55.780353 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"093e52d3-1df3-49de-97a6-bd8f08bf7a8f","Type":"ContainerDied","Data":"ac4d6c273d54bed796cd6768d8b63a82b838e4a3bb38b86d96f31bdb5233d409"} Dec 02 01:56:55 crc kubenswrapper[4884]: I1202 01:56:55.782638 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fb5702eb-cee9-4049-b3b4-601f8f5e1854","Type":"ContainerStarted","Data":"06a8b59e950e9d35a57f05254d11958b7f2b55b279c3bdc1ddcdbbebfe783b2c"} Dec 02 01:56:55 crc kubenswrapper[4884]: I1202 01:56:55.786246 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5fdad2f7-f276-458e-96dd-6cc7fda79cac","Type":"ContainerStarted","Data":"40818d37cad862df2661fcdde77ac0513e9e9b1c5f7701e916bdaee848403a76"} Dec 02 01:56:55 crc kubenswrapper[4884]: I1202 01:56:55.812518 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.812500799 podStartE2EDuration="8.812500799s" podCreationTimestamp="2025-12-02 01:56:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:55.80441053 +0000 UTC m=+1232.480247424" watchObservedRunningTime="2025-12-02 01:56:55.812500799 +0000 UTC m=+1232.488337683" Dec 02 01:56:56 crc kubenswrapper[4884]: I1202 01:56:56.803926 4884 generic.go:334] "Generic (PLEG): container finished" podID="fe8c221e-e320-4358-b940-436f38a423d3" containerID="1989683bf093eb5ecb39497dbe41e94fa661c162853fe0b1a5047b3edb0cbeb1" exitCode=0 Dec 02 01:56:56 crc kubenswrapper[4884]: I1202 01:56:56.804836 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wmhnq" event={"ID":"fe8c221e-e320-4358-b940-436f38a423d3","Type":"ContainerDied","Data":"1989683bf093eb5ecb39497dbe41e94fa661c162853fe0b1a5047b3edb0cbeb1"} Dec 02 01:56:57 crc kubenswrapper[4884]: I1202 01:56:57.815267 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fb5702eb-cee9-4049-b3b4-601f8f5e1854","Type":"ContainerStarted","Data":"6448129f68147a08fa73ca4e4a2c9db98fbe699e03fcc1e147e3aa27b7eff2ec"} Dec 02 01:56:57 crc kubenswrapper[4884]: I1202 01:56:57.817058 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-c9rsj" event={"ID":"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada","Type":"ContainerStarted","Data":"ea22b91c2dde8cd499c7ebeb771560a39adc369d8ad7edd3dfb5d0effe098cd4"} Dec 02 01:56:57 crc kubenswrapper[4884]: I1202 01:56:57.822813 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45","Type":"ContainerStarted","Data":"e5249e12843a0b5a923b2d8cbcb00cd0581382a128766e190f68ffd31688caf4"} Dec 02 01:56:57 crc kubenswrapper[4884]: I1202 01:56:57.842760 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-c9rsj" podStartSLOduration=2.763900997 podStartE2EDuration="48.842728364s" podCreationTimestamp="2025-12-02 01:56:09 +0000 UTC" firstStartedPulling="2025-12-02 01:56:11.08113216 +0000 UTC m=+1187.756969044" lastFinishedPulling="2025-12-02 01:56:57.159959527 +0000 UTC m=+1233.835796411" observedRunningTime="2025-12-02 01:56:57.839074934 +0000 UTC m=+1234.514911828" watchObservedRunningTime="2025-12-02 01:56:57.842728364 +0000 UTC m=+1234.518565248" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.155992 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="093e52d3-1df3-49de-97a6-bd8f08bf7a8f" containerName="watcher-api-log" probeResult="failure" output="Get \"http://10.217.0.164:9322/\": read tcp 10.217.0.2:35854->10.217.0.164:9322: read: connection reset by peer" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.156043 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="093e52d3-1df3-49de-97a6-bd8f08bf7a8f" containerName="watcher-api" probeResult="failure" output="Get \"http://10.217.0.164:9322/\": read tcp 10.217.0.2:35846->10.217.0.164:9322: read: connection reset by peer" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.310589 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.310645 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.320625 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wmhnq" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.373901 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.380176 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.406564 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-785d8bcb8c-crspk" podUID="4f6fe6c3-c15e-45d3-91e5-1474030f2bbd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.155:5353: i/o timeout" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.452781 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fe8c221e-e320-4358-b940-436f38a423d3-db-sync-config-data\") pod \"fe8c221e-e320-4358-b940-436f38a423d3\" (UID: \"fe8c221e-e320-4358-b940-436f38a423d3\") " Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.452895 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe8c221e-e320-4358-b940-436f38a423d3-combined-ca-bundle\") pod \"fe8c221e-e320-4358-b940-436f38a423d3\" (UID: \"fe8c221e-e320-4358-b940-436f38a423d3\") " Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.452993 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjdq9\" (UniqueName: \"kubernetes.io/projected/fe8c221e-e320-4358-b940-436f38a423d3-kube-api-access-pjdq9\") pod \"fe8c221e-e320-4358-b940-436f38a423d3\" (UID: \"fe8c221e-e320-4358-b940-436f38a423d3\") " Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.458807 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe8c221e-e320-4358-b940-436f38a423d3-kube-api-access-pjdq9" (OuterVolumeSpecName: "kube-api-access-pjdq9") pod "fe8c221e-e320-4358-b940-436f38a423d3" (UID: "fe8c221e-e320-4358-b940-436f38a423d3"). InnerVolumeSpecName "kube-api-access-pjdq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.468229 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe8c221e-e320-4358-b940-436f38a423d3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "fe8c221e-e320-4358-b940-436f38a423d3" (UID: "fe8c221e-e320-4358-b940-436f38a423d3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.498516 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe8c221e-e320-4358-b940-436f38a423d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe8c221e-e320-4358-b940-436f38a423d3" (UID: "fe8c221e-e320-4358-b940-436f38a423d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.555929 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjdq9\" (UniqueName: \"kubernetes.io/projected/fe8c221e-e320-4358-b940-436f38a423d3-kube-api-access-pjdq9\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.555962 4884 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/fe8c221e-e320-4358-b940-436f38a423d3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.555974 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe8c221e-e320-4358-b940-436f38a423d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.591696 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.760393 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-config-data\") pod \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.760793 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-custom-prometheus-ca\") pod \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.761206 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-logs\") pod \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.761274 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-combined-ca-bundle\") pod \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.761315 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlzhl\" (UniqueName: \"kubernetes.io/projected/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-kube-api-access-wlzhl\") pod \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\" (UID: \"093e52d3-1df3-49de-97a6-bd8f08bf7a8f\") " Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.765081 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-logs" (OuterVolumeSpecName: "logs") pod "093e52d3-1df3-49de-97a6-bd8f08bf7a8f" (UID: "093e52d3-1df3-49de-97a6-bd8f08bf7a8f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.776231 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-kube-api-access-wlzhl" (OuterVolumeSpecName: "kube-api-access-wlzhl") pod "093e52d3-1df3-49de-97a6-bd8f08bf7a8f" (UID: "093e52d3-1df3-49de-97a6-bd8f08bf7a8f"). InnerVolumeSpecName "kube-api-access-wlzhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.811152 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "093e52d3-1df3-49de-97a6-bd8f08bf7a8f" (UID: "093e52d3-1df3-49de-97a6-bd8f08bf7a8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.811703 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "093e52d3-1df3-49de-97a6-bd8f08bf7a8f" (UID: "093e52d3-1df3-49de-97a6-bd8f08bf7a8f"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.821216 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-config-data" (OuterVolumeSpecName: "config-data") pod "093e52d3-1df3-49de-97a6-bd8f08bf7a8f" (UID: "093e52d3-1df3-49de-97a6-bd8f08bf7a8f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.838092 4884 generic.go:334] "Generic (PLEG): container finished" podID="093e52d3-1df3-49de-97a6-bd8f08bf7a8f" containerID="365c3fe6e851399da4164648cffe6fd37fbdc9a8927b1644bf300c0856806c26" exitCode=0 Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.838155 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.838153 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"093e52d3-1df3-49de-97a6-bd8f08bf7a8f","Type":"ContainerDied","Data":"365c3fe6e851399da4164648cffe6fd37fbdc9a8927b1644bf300c0856806c26"} Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.838264 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"093e52d3-1df3-49de-97a6-bd8f08bf7a8f","Type":"ContainerDied","Data":"5d62346cada22f97af6bc99b8d3c1377739ca065c83cea2cf5bc6ee3cc09906f"} Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.838284 4884 scope.go:117] "RemoveContainer" containerID="365c3fe6e851399da4164648cffe6fd37fbdc9a8927b1644bf300c0856806c26" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.842827 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fb5702eb-cee9-4049-b3b4-601f8f5e1854","Type":"ContainerStarted","Data":"d437543659450b227db4b3f26ac8d8eeea74224d10402b9ccfcd8185c9ae56be"} Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.855429 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wmhnq" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.855480 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wmhnq" event={"ID":"fe8c221e-e320-4358-b940-436f38a423d3","Type":"ContainerDied","Data":"7c3d69fd23b6cc1507a9b34b533179bac25cd94f108362939294514e1ab412f1"} Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.855509 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c3d69fd23b6cc1507a9b34b533179bac25cd94f108362939294514e1ab412f1" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.856248 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.856272 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.864442 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.864960 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlzhl\" (UniqueName: \"kubernetes.io/projected/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-kube-api-access-wlzhl\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.865127 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.865290 4884 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.865409 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/093e52d3-1df3-49de-97a6-bd8f08bf7a8f-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.892567 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.892545762 podStartE2EDuration="5.892545762s" podCreationTimestamp="2025-12-02 01:56:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:56:58.869998865 +0000 UTC m=+1235.545835759" watchObservedRunningTime="2025-12-02 01:56:58.892545762 +0000 UTC m=+1235.568382646" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.906607 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-api-0"] Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.914088 4884 scope.go:117] "RemoveContainer" containerID="ac4d6c273d54bed796cd6768d8b63a82b838e4a3bb38b86d96f31bdb5233d409" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.928045 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-api-0"] Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.933843 4884 scope.go:117] "RemoveContainer" containerID="365c3fe6e851399da4164648cffe6fd37fbdc9a8927b1644bf300c0856806c26" Dec 02 01:56:58 crc kubenswrapper[4884]: E1202 01:56:58.936567 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"365c3fe6e851399da4164648cffe6fd37fbdc9a8927b1644bf300c0856806c26\": container with ID starting with 365c3fe6e851399da4164648cffe6fd37fbdc9a8927b1644bf300c0856806c26 not found: ID does not exist" containerID="365c3fe6e851399da4164648cffe6fd37fbdc9a8927b1644bf300c0856806c26" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.936605 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"365c3fe6e851399da4164648cffe6fd37fbdc9a8927b1644bf300c0856806c26"} err="failed to get container status \"365c3fe6e851399da4164648cffe6fd37fbdc9a8927b1644bf300c0856806c26\": rpc error: code = NotFound desc = could not find container \"365c3fe6e851399da4164648cffe6fd37fbdc9a8927b1644bf300c0856806c26\": container with ID starting with 365c3fe6e851399da4164648cffe6fd37fbdc9a8927b1644bf300c0856806c26 not found: ID does not exist" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.936627 4884 scope.go:117] "RemoveContainer" containerID="ac4d6c273d54bed796cd6768d8b63a82b838e4a3bb38b86d96f31bdb5233d409" Dec 02 01:56:58 crc kubenswrapper[4884]: E1202 01:56:58.937811 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac4d6c273d54bed796cd6768d8b63a82b838e4a3bb38b86d96f31bdb5233d409\": container with ID starting with ac4d6c273d54bed796cd6768d8b63a82b838e4a3bb38b86d96f31bdb5233d409 not found: ID does not exist" containerID="ac4d6c273d54bed796cd6768d8b63a82b838e4a3bb38b86d96f31bdb5233d409" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.937833 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac4d6c273d54bed796cd6768d8b63a82b838e4a3bb38b86d96f31bdb5233d409"} err="failed to get container status \"ac4d6c273d54bed796cd6768d8b63a82b838e4a3bb38b86d96f31bdb5233d409\": rpc error: code = NotFound desc = could not find container \"ac4d6c273d54bed796cd6768d8b63a82b838e4a3bb38b86d96f31bdb5233d409\": container with ID starting with ac4d6c273d54bed796cd6768d8b63a82b838e4a3bb38b86d96f31bdb5233d409 not found: ID does not exist" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.964821 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-api-0"] Dec 02 01:56:58 crc kubenswrapper[4884]: E1202 01:56:58.965229 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093e52d3-1df3-49de-97a6-bd8f08bf7a8f" containerName="watcher-api-log" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.965250 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="093e52d3-1df3-49de-97a6-bd8f08bf7a8f" containerName="watcher-api-log" Dec 02 01:56:58 crc kubenswrapper[4884]: E1202 01:56:58.965265 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe8c221e-e320-4358-b940-436f38a423d3" containerName="barbican-db-sync" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.965273 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe8c221e-e320-4358-b940-436f38a423d3" containerName="barbican-db-sync" Dec 02 01:56:58 crc kubenswrapper[4884]: E1202 01:56:58.965283 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093e52d3-1df3-49de-97a6-bd8f08bf7a8f" containerName="watcher-api" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.965290 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="093e52d3-1df3-49de-97a6-bd8f08bf7a8f" containerName="watcher-api" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.965491 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="093e52d3-1df3-49de-97a6-bd8f08bf7a8f" containerName="watcher-api-log" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.965512 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe8c221e-e320-4358-b940-436f38a423d3" containerName="barbican-db-sync" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.965528 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="093e52d3-1df3-49de-97a6-bd8f08bf7a8f" containerName="watcher-api" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.966478 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.993821 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-api-config-data" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.994114 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-public-svc" Dec 02 01:56:58 crc kubenswrapper[4884]: I1202 01:56:58.994616 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-watcher-internal-svc" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.001832 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.101570 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a84d6c24-16a7-43a2-a303-5befb347c8f4-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.130913 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a84d6c24-16a7-43a2-a303-5befb347c8f4-config-data\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.131172 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a84d6c24-16a7-43a2-a303-5befb347c8f4-public-tls-certs\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.131257 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a84d6c24-16a7-43a2-a303-5befb347c8f4-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.131379 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8vj4\" (UniqueName: \"kubernetes.io/projected/a84d6c24-16a7-43a2-a303-5befb347c8f4-kube-api-access-g8vj4\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.131496 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a84d6c24-16a7-43a2-a303-5befb347c8f4-logs\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.131621 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a84d6c24-16a7-43a2-a303-5befb347c8f4-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.204678 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7b4498cc99-xx9bb"] Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.215383 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.222825 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7b4498cc99-xx9bb"] Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.226904 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-55fc6c9b58-9v749" podUID="59b51ab4-e0c8-42af-92aa-063e66157aa7" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.159:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.159:8443: connect: connection refused" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.228073 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-lmbq5" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.228280 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.228514 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.243698 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a84d6c24-16a7-43a2-a303-5befb347c8f4-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.249459 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a84d6c24-16a7-43a2-a303-5befb347c8f4-config-data\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.249639 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a84d6c24-16a7-43a2-a303-5befb347c8f4-public-tls-certs\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.249719 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a84d6c24-16a7-43a2-a303-5befb347c8f4-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.249895 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8vj4\" (UniqueName: \"kubernetes.io/projected/a84d6c24-16a7-43a2-a303-5befb347c8f4-kube-api-access-g8vj4\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.250014 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a84d6c24-16a7-43a2-a303-5befb347c8f4-logs\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.250406 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a84d6c24-16a7-43a2-a303-5befb347c8f4-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.252939 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-8687f948dd-5zg8m" podUID="19afec6c-37bd-4e1a-9f96-c7e37c1e7748" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.160:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.160:8443: connect: connection refused" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.254015 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/a84d6c24-16a7-43a2-a303-5befb347c8f4-custom-prometheus-ca\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.254794 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a84d6c24-16a7-43a2-a303-5befb347c8f4-logs\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.254820 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-57b6695dcb-8c27t"] Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.257635 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.274593 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a84d6c24-16a7-43a2-a303-5befb347c8f4-config-data\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.276223 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a84d6c24-16a7-43a2-a303-5befb347c8f4-public-tls-certs\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.280141 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.285367 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a84d6c24-16a7-43a2-a303-5befb347c8f4-internal-tls-certs\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.285907 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a84d6c24-16a7-43a2-a303-5befb347c8f4-combined-ca-bundle\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.308212 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8vj4\" (UniqueName: \"kubernetes.io/projected/a84d6c24-16a7-43a2-a303-5befb347c8f4-kube-api-access-g8vj4\") pod \"watcher-api-0\" (UID: \"a84d6c24-16a7-43a2-a303-5befb347c8f4\") " pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.314336 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-57b6695dcb-8c27t"] Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.336110 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-api-0" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.351702 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0163d38-d25f-460e-b563-7d7198b52999-config-data\") pod \"barbican-keystone-listener-57b6695dcb-8c27t\" (UID: \"e0163d38-d25f-460e-b563-7d7198b52999\") " pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.351854 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0163d38-d25f-460e-b563-7d7198b52999-config-data-custom\") pod \"barbican-keystone-listener-57b6695dcb-8c27t\" (UID: \"e0163d38-d25f-460e-b563-7d7198b52999\") " pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.351890 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxgx8\" (UniqueName: \"kubernetes.io/projected/e0163d38-d25f-460e-b563-7d7198b52999-kube-api-access-jxgx8\") pod \"barbican-keystone-listener-57b6695dcb-8c27t\" (UID: \"e0163d38-d25f-460e-b563-7d7198b52999\") " pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.351910 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0163d38-d25f-460e-b563-7d7198b52999-logs\") pod \"barbican-keystone-listener-57b6695dcb-8c27t\" (UID: \"e0163d38-d25f-460e-b563-7d7198b52999\") " pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.351941 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lxkd\" (UniqueName: \"kubernetes.io/projected/80449fc3-b0b7-45fd-9027-49ca53a34bb6-kube-api-access-6lxkd\") pod \"barbican-worker-7b4498cc99-xx9bb\" (UID: \"80449fc3-b0b7-45fd-9027-49ca53a34bb6\") " pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.351960 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80449fc3-b0b7-45fd-9027-49ca53a34bb6-config-data\") pod \"barbican-worker-7b4498cc99-xx9bb\" (UID: \"80449fc3-b0b7-45fd-9027-49ca53a34bb6\") " pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.351977 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0163d38-d25f-460e-b563-7d7198b52999-combined-ca-bundle\") pod \"barbican-keystone-listener-57b6695dcb-8c27t\" (UID: \"e0163d38-d25f-460e-b563-7d7198b52999\") " pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.351991 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80449fc3-b0b7-45fd-9027-49ca53a34bb6-config-data-custom\") pod \"barbican-worker-7b4498cc99-xx9bb\" (UID: \"80449fc3-b0b7-45fd-9027-49ca53a34bb6\") " pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.352007 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80449fc3-b0b7-45fd-9027-49ca53a34bb6-logs\") pod \"barbican-worker-7b4498cc99-xx9bb\" (UID: \"80449fc3-b0b7-45fd-9027-49ca53a34bb6\") " pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.352055 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80449fc3-b0b7-45fd-9027-49ca53a34bb6-combined-ca-bundle\") pod \"barbican-worker-7b4498cc99-xx9bb\" (UID: \"80449fc3-b0b7-45fd-9027-49ca53a34bb6\") " pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.407085 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-2j759"] Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.408651 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.431495 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-2j759"] Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.448044 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6588b54d58-b7m95"] Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.449514 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.467582 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.471855 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80449fc3-b0b7-45fd-9027-49ca53a34bb6-config-data\") pod \"barbican-worker-7b4498cc99-xx9bb\" (UID: \"80449fc3-b0b7-45fd-9027-49ca53a34bb6\") " pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.471909 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0163d38-d25f-460e-b563-7d7198b52999-combined-ca-bundle\") pod \"barbican-keystone-listener-57b6695dcb-8c27t\" (UID: \"e0163d38-d25f-460e-b563-7d7198b52999\") " pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.471933 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80449fc3-b0b7-45fd-9027-49ca53a34bb6-config-data-custom\") pod \"barbican-worker-7b4498cc99-xx9bb\" (UID: \"80449fc3-b0b7-45fd-9027-49ca53a34bb6\") " pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.471950 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80449fc3-b0b7-45fd-9027-49ca53a34bb6-logs\") pod \"barbican-worker-7b4498cc99-xx9bb\" (UID: \"80449fc3-b0b7-45fd-9027-49ca53a34bb6\") " pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.471972 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.472011 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgfj8\" (UniqueName: \"kubernetes.io/projected/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-kube-api-access-fgfj8\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.472091 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80449fc3-b0b7-45fd-9027-49ca53a34bb6-combined-ca-bundle\") pod \"barbican-worker-7b4498cc99-xx9bb\" (UID: \"80449fc3-b0b7-45fd-9027-49ca53a34bb6\") " pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.472171 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0163d38-d25f-460e-b563-7d7198b52999-config-data\") pod \"barbican-keystone-listener-57b6695dcb-8c27t\" (UID: \"e0163d38-d25f-460e-b563-7d7198b52999\") " pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.472192 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-config\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.472206 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-dns-svc\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.472232 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.472298 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0163d38-d25f-460e-b563-7d7198b52999-config-data-custom\") pod \"barbican-keystone-listener-57b6695dcb-8c27t\" (UID: \"e0163d38-d25f-460e-b563-7d7198b52999\") " pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.472329 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxgx8\" (UniqueName: \"kubernetes.io/projected/e0163d38-d25f-460e-b563-7d7198b52999-kube-api-access-jxgx8\") pod \"barbican-keystone-listener-57b6695dcb-8c27t\" (UID: \"e0163d38-d25f-460e-b563-7d7198b52999\") " pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.472354 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0163d38-d25f-460e-b563-7d7198b52999-logs\") pod \"barbican-keystone-listener-57b6695dcb-8c27t\" (UID: \"e0163d38-d25f-460e-b563-7d7198b52999\") " pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.472389 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.472442 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lxkd\" (UniqueName: \"kubernetes.io/projected/80449fc3-b0b7-45fd-9027-49ca53a34bb6-kube-api-access-6lxkd\") pod \"barbican-worker-7b4498cc99-xx9bb\" (UID: \"80449fc3-b0b7-45fd-9027-49ca53a34bb6\") " pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.472997 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0163d38-d25f-460e-b563-7d7198b52999-logs\") pod \"barbican-keystone-listener-57b6695dcb-8c27t\" (UID: \"e0163d38-d25f-460e-b563-7d7198b52999\") " pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.476676 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80449fc3-b0b7-45fd-9027-49ca53a34bb6-logs\") pod \"barbican-worker-7b4498cc99-xx9bb\" (UID: \"80449fc3-b0b7-45fd-9027-49ca53a34bb6\") " pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.483878 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80449fc3-b0b7-45fd-9027-49ca53a34bb6-config-data\") pod \"barbican-worker-7b4498cc99-xx9bb\" (UID: \"80449fc3-b0b7-45fd-9027-49ca53a34bb6\") " pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.487139 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80449fc3-b0b7-45fd-9027-49ca53a34bb6-config-data-custom\") pod \"barbican-worker-7b4498cc99-xx9bb\" (UID: \"80449fc3-b0b7-45fd-9027-49ca53a34bb6\") " pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.490804 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6588b54d58-b7m95"] Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.501543 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0163d38-d25f-460e-b563-7d7198b52999-combined-ca-bundle\") pod \"barbican-keystone-listener-57b6695dcb-8c27t\" (UID: \"e0163d38-d25f-460e-b563-7d7198b52999\") " pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.504543 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80449fc3-b0b7-45fd-9027-49ca53a34bb6-combined-ca-bundle\") pod \"barbican-worker-7b4498cc99-xx9bb\" (UID: \"80449fc3-b0b7-45fd-9027-49ca53a34bb6\") " pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.511040 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lxkd\" (UniqueName: \"kubernetes.io/projected/80449fc3-b0b7-45fd-9027-49ca53a34bb6-kube-api-access-6lxkd\") pod \"barbican-worker-7b4498cc99-xx9bb\" (UID: \"80449fc3-b0b7-45fd-9027-49ca53a34bb6\") " pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.515908 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e0163d38-d25f-460e-b563-7d7198b52999-config-data-custom\") pod \"barbican-keystone-listener-57b6695dcb-8c27t\" (UID: \"e0163d38-d25f-460e-b563-7d7198b52999\") " pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.515947 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxgx8\" (UniqueName: \"kubernetes.io/projected/e0163d38-d25f-460e-b563-7d7198b52999-kube-api-access-jxgx8\") pod \"barbican-keystone-listener-57b6695dcb-8c27t\" (UID: \"e0163d38-d25f-460e-b563-7d7198b52999\") " pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.520595 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0163d38-d25f-460e-b563-7d7198b52999-config-data\") pod \"barbican-keystone-listener-57b6695dcb-8c27t\" (UID: \"e0163d38-d25f-460e-b563-7d7198b52999\") " pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.573548 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-combined-ca-bundle\") pod \"barbican-api-6588b54d58-b7m95\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.573597 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.573646 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.573669 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-config-data\") pod \"barbican-api-6588b54d58-b7m95\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.573688 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgfj8\" (UniqueName: \"kubernetes.io/projected/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-kube-api-access-fgfj8\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.573733 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-config-data-custom\") pod \"barbican-api-6588b54d58-b7m95\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.573784 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-logs\") pod \"barbican-api-6588b54d58-b7m95\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.573801 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpkz8\" (UniqueName: \"kubernetes.io/projected/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-kube-api-access-hpkz8\") pod \"barbican-api-6588b54d58-b7m95\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.573828 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-config\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.573843 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-dns-svc\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.573864 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.574698 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.575220 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.576185 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.576192 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-config\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.576891 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-dns-svc\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.601705 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgfj8\" (UniqueName: \"kubernetes.io/projected/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-kube-api-access-fgfj8\") pod \"dnsmasq-dns-85ff748b95-2j759\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.633525 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="093e52d3-1df3-49de-97a6-bd8f08bf7a8f" path="/var/lib/kubelet/pods/093e52d3-1df3-49de-97a6-bd8f08bf7a8f/volumes" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.675413 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7b4498cc99-xx9bb" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.675862 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-config-data\") pod \"barbican-api-6588b54d58-b7m95\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.675963 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-config-data-custom\") pod \"barbican-api-6588b54d58-b7m95\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.676007 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-logs\") pod \"barbican-api-6588b54d58-b7m95\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.676025 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpkz8\" (UniqueName: \"kubernetes.io/projected/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-kube-api-access-hpkz8\") pod \"barbican-api-6588b54d58-b7m95\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.676131 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-combined-ca-bundle\") pod \"barbican-api-6588b54d58-b7m95\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.676684 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-logs\") pod \"barbican-api-6588b54d58-b7m95\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.682842 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-config-data-custom\") pod \"barbican-api-6588b54d58-b7m95\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.688250 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-combined-ca-bundle\") pod \"barbican-api-6588b54d58-b7m95\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.691462 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-config-data\") pod \"barbican-api-6588b54d58-b7m95\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.700331 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpkz8\" (UniqueName: \"kubernetes.io/projected/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-kube-api-access-hpkz8\") pod \"barbican-api-6588b54d58-b7m95\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.802296 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.832027 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.844574 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:56:59 crc kubenswrapper[4884]: I1202 01:56:59.994649 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-api-0"] Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.262860 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7b4498cc99-xx9bb"] Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.404296 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-2j759"] Dec 02 01:57:00 crc kubenswrapper[4884]: W1202 01:57:00.432079 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a4e2ac4_dc7b_46f4_bee0_7e6670b07990.slice/crio-00170689d0d5774d471c147bf0e58ff1e9ac8799f0a93b4f73ebffb783b102de WatchSource:0}: Error finding container 00170689d0d5774d471c147bf0e58ff1e9ac8799f0a93b4f73ebffb783b102de: Status 404 returned error can't find the container with id 00170689d0d5774d471c147bf0e58ff1e9ac8799f0a93b4f73ebffb783b102de Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.557733 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6588b54d58-b7m95"] Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.578121 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-57b6695dcb-8c27t"] Dec 02 01:57:00 crc kubenswrapper[4884]: W1202 01:57:00.587880 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0163d38_d25f_460e_b563_7d7198b52999.slice/crio-4d5d276d40771a9e1109208d659ccf67c1f5f97febef039976099057daa0bb59 WatchSource:0}: Error finding container 4d5d276d40771a9e1109208d659ccf67c1f5f97febef039976099057daa0bb59: Status 404 returned error can't find the container with id 4d5d276d40771a9e1109208d659ccf67c1f5f97febef039976099057daa0bb59 Dec 02 01:57:00 crc kubenswrapper[4884]: W1202 01:57:00.590097 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4536243_ac9e_4f38_ac2f_a8649e3b4b22.slice/crio-97cad8e2c94ff5bed01a28844faa38aae1e69a1c88b142d2a0b8bf16ff38ac84 WatchSource:0}: Error finding container 97cad8e2c94ff5bed01a28844faa38aae1e69a1c88b142d2a0b8bf16ff38ac84: Status 404 returned error can't find the container with id 97cad8e2c94ff5bed01a28844faa38aae1e69a1c88b142d2a0b8bf16ff38ac84 Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.903133 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" event={"ID":"e0163d38-d25f-460e-b563-7d7198b52999","Type":"ContainerStarted","Data":"4d5d276d40771a9e1109208d659ccf67c1f5f97febef039976099057daa0bb59"} Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.909690 4884 generic.go:334] "Generic (PLEG): container finished" podID="9a4e2ac4-dc7b-46f4-bee0-7e6670b07990" containerID="3cf5e61393f4ec424c6be9f60d610c754eaf902dfa8813b3442375974ac4945f" exitCode=0 Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.909736 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-2j759" event={"ID":"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990","Type":"ContainerDied","Data":"3cf5e61393f4ec424c6be9f60d610c754eaf902dfa8813b3442375974ac4945f"} Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.909785 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-2j759" event={"ID":"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990","Type":"ContainerStarted","Data":"00170689d0d5774d471c147bf0e58ff1e9ac8799f0a93b4f73ebffb783b102de"} Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.915562 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b4498cc99-xx9bb" event={"ID":"80449fc3-b0b7-45fd-9027-49ca53a34bb6","Type":"ContainerStarted","Data":"7f267443319feccc13ec480e2511fed6eb028bee1d16eeecfe068ca15511223f"} Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.918242 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"a84d6c24-16a7-43a2-a303-5befb347c8f4","Type":"ContainerStarted","Data":"506c6dc8f71aae614f9f7f17ce50ebc377a219d79b1cef701f359e5db0686249"} Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.918274 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"a84d6c24-16a7-43a2-a303-5befb347c8f4","Type":"ContainerStarted","Data":"a15365fe7efc733ab512859b60244b4cd1a650d45cc3dad0a8d97f32ec935e98"} Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.918285 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-api-0" event={"ID":"a84d6c24-16a7-43a2-a303-5befb347c8f4","Type":"ContainerStarted","Data":"ce21db19388dce2917f1c36f49d2fb24b665051cd72c4a0de20bcbb637e73105"} Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.920623 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.930040 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/watcher-api-0" podUID="a84d6c24-16a7-43a2-a303-5befb347c8f4" containerName="watcher-api" probeResult="failure" output="Get \"https://10.217.0.171:9322/\": dial tcp 10.217.0.171:9322: connect: connection refused" Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.930638 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6588b54d58-b7m95" event={"ID":"e4536243-ac9e-4f38-ac2f-a8649e3b4b22","Type":"ContainerStarted","Data":"208eb2afa2b522caa1da161847bde9c2d7cee195285c683fa02623a06b368fdc"} Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.930674 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6588b54d58-b7m95" event={"ID":"e4536243-ac9e-4f38-ac2f-a8649e3b4b22","Type":"ContainerStarted","Data":"97cad8e2c94ff5bed01a28844faa38aae1e69a1c88b142d2a0b8bf16ff38ac84"} Dec 02 01:57:00 crc kubenswrapper[4884]: I1202 01:57:00.950910 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-api-0" podStartSLOduration=2.950892071 podStartE2EDuration="2.950892071s" podCreationTimestamp="2025-12-02 01:56:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:57:00.948907882 +0000 UTC m=+1237.624744766" watchObservedRunningTime="2025-12-02 01:57:00.950892071 +0000 UTC m=+1237.626728945" Dec 02 01:57:01 crc kubenswrapper[4884]: I1202 01:57:01.945553 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-2j759" event={"ID":"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990","Type":"ContainerStarted","Data":"74dfaa214b6c87de35bf84429f2cc26ce9f16580512e52fc422389f7f16e4e06"} Dec 02 01:57:01 crc kubenswrapper[4884]: I1202 01:57:01.947610 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:57:01 crc kubenswrapper[4884]: I1202 01:57:01.953804 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6588b54d58-b7m95" event={"ID":"e4536243-ac9e-4f38-ac2f-a8649e3b4b22","Type":"ContainerStarted","Data":"2182bc11bb7a544dd3b8deee688205b5ef124cc4136a5fc9e6c6ee486e7e245f"} Dec 02 01:57:01 crc kubenswrapper[4884]: I1202 01:57:01.954161 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:57:01 crc kubenswrapper[4884]: I1202 01:57:01.980026 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-2j759" podStartSLOduration=2.980007199 podStartE2EDuration="2.980007199s" podCreationTimestamp="2025-12-02 01:56:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:57:01.971732195 +0000 UTC m=+1238.647569089" watchObservedRunningTime="2025-12-02 01:57:01.980007199 +0000 UTC m=+1238.655844083" Dec 02 01:57:01 crc kubenswrapper[4884]: I1202 01:57:01.998970 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6588b54d58-b7m95" podStartSLOduration=2.998952597 podStartE2EDuration="2.998952597s" podCreationTimestamp="2025-12-02 01:56:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:57:01.998107116 +0000 UTC m=+1238.673944020" watchObservedRunningTime="2025-12-02 01:57:01.998952597 +0000 UTC m=+1238.674789491" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.546516 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5fb984d75d-76xfp"] Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.548066 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.555453 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.555938 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.570300 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5fb984d75d-76xfp"] Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.668752 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6059a81d-e008-4cba-8af1-9b093989ddf2-config-data-custom\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.669096 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6059a81d-e008-4cba-8af1-9b093989ddf2-combined-ca-bundle\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.669166 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6059a81d-e008-4cba-8af1-9b093989ddf2-logs\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.669246 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6wdh\" (UniqueName: \"kubernetes.io/projected/6059a81d-e008-4cba-8af1-9b093989ddf2-kube-api-access-r6wdh\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.669267 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6059a81d-e008-4cba-8af1-9b093989ddf2-config-data\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.669307 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6059a81d-e008-4cba-8af1-9b093989ddf2-public-tls-certs\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.669538 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6059a81d-e008-4cba-8af1-9b093989ddf2-internal-tls-certs\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.736095 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.736625 4884 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.748172 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.770864 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6059a81d-e008-4cba-8af1-9b093989ddf2-internal-tls-certs\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.770911 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6059a81d-e008-4cba-8af1-9b093989ddf2-config-data-custom\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.770961 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6059a81d-e008-4cba-8af1-9b093989ddf2-combined-ca-bundle\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.771019 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6059a81d-e008-4cba-8af1-9b093989ddf2-logs\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.771083 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6wdh\" (UniqueName: \"kubernetes.io/projected/6059a81d-e008-4cba-8af1-9b093989ddf2-kube-api-access-r6wdh\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.771101 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6059a81d-e008-4cba-8af1-9b093989ddf2-config-data\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.771121 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6059a81d-e008-4cba-8af1-9b093989ddf2-public-tls-certs\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.772077 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6059a81d-e008-4cba-8af1-9b093989ddf2-logs\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.777863 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6059a81d-e008-4cba-8af1-9b093989ddf2-internal-tls-certs\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.778847 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6059a81d-e008-4cba-8af1-9b093989ddf2-combined-ca-bundle\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.779403 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6059a81d-e008-4cba-8af1-9b093989ddf2-config-data-custom\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.780476 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6059a81d-e008-4cba-8af1-9b093989ddf2-config-data\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.783036 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6059a81d-e008-4cba-8af1-9b093989ddf2-public-tls-certs\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.790307 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6wdh\" (UniqueName: \"kubernetes.io/projected/6059a81d-e008-4cba-8af1-9b093989ddf2-kube-api-access-r6wdh\") pod \"barbican-api-5fb984d75d-76xfp\" (UID: \"6059a81d-e008-4cba-8af1-9b093989ddf2\") " pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.872604 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:02 crc kubenswrapper[4884]: I1202 01:57:02.967449 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:57:03 crc kubenswrapper[4884]: I1202 01:57:03.980317 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" event={"ID":"e0163d38-d25f-460e-b563-7d7198b52999","Type":"ContainerStarted","Data":"9574be78d219666a63b321986a09b085ecf2795316bd522b4e6c74a1b0e2c283"} Dec 02 01:57:03 crc kubenswrapper[4884]: I1202 01:57:03.986328 4884 generic.go:334] "Generic (PLEG): container finished" podID="f0d68bd4-6c59-4e41-b32e-f7918d3b4ada" containerID="ea22b91c2dde8cd499c7ebeb771560a39adc369d8ad7edd3dfb5d0effe098cd4" exitCode=0 Dec 02 01:57:03 crc kubenswrapper[4884]: I1202 01:57:03.986376 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-c9rsj" event={"ID":"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada","Type":"ContainerDied","Data":"ea22b91c2dde8cd499c7ebeb771560a39adc369d8ad7edd3dfb5d0effe098cd4"} Dec 02 01:57:04 crc kubenswrapper[4884]: I1202 01:57:04.003476 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b4498cc99-xx9bb" event={"ID":"80449fc3-b0b7-45fd-9027-49ca53a34bb6","Type":"ContainerStarted","Data":"0ac394570905ef6502e23f33b9a2c20c9514191e2a0a38efb02cf052306965b6"} Dec 02 01:57:04 crc kubenswrapper[4884]: I1202 01:57:04.011591 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5fb984d75d-76xfp"] Dec 02 01:57:04 crc kubenswrapper[4884]: I1202 01:57:04.188463 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 02 01:57:04 crc kubenswrapper[4884]: I1202 01:57:04.189423 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 02 01:57:04 crc kubenswrapper[4884]: I1202 01:57:04.224110 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 02 01:57:04 crc kubenswrapper[4884]: I1202 01:57:04.238358 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 02 01:57:04 crc kubenswrapper[4884]: I1202 01:57:04.336959 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-api-0" Dec 02 01:57:04 crc kubenswrapper[4884]: I1202 01:57:04.337069 4884 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 01:57:04 crc kubenswrapper[4884]: I1202 01:57:04.758248 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Dec 02 01:57:05 crc kubenswrapper[4884]: I1202 01:57:05.015424 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fb984d75d-76xfp" event={"ID":"6059a81d-e008-4cba-8af1-9b093989ddf2","Type":"ContainerStarted","Data":"1f1e307e19eaa555b6aa8435ffb954b9d5084d8d5330ee1ffa9ccfa082342cd5"} Dec 02 01:57:05 crc kubenswrapper[4884]: I1202 01:57:05.015487 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fb984d75d-76xfp" event={"ID":"6059a81d-e008-4cba-8af1-9b093989ddf2","Type":"ContainerStarted","Data":"f0dc10bd33c44cea3255245c6240c2bf3f74a7fa048320b7d878e5dda172a984"} Dec 02 01:57:05 crc kubenswrapper[4884]: I1202 01:57:05.021670 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" event={"ID":"e0163d38-d25f-460e-b563-7d7198b52999","Type":"ContainerStarted","Data":"31ac3a951a8912ec2a877ff802c8c0c75c7fd83b736e40eb949e2578fab34d61"} Dec 02 01:57:05 crc kubenswrapper[4884]: I1202 01:57:05.030812 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b4498cc99-xx9bb" event={"ID":"80449fc3-b0b7-45fd-9027-49ca53a34bb6","Type":"ContainerStarted","Data":"907eb0cf5b51cee989ba3101ebf1415fc581adeee33af2f9a564897c1406b35f"} Dec 02 01:57:05 crc kubenswrapper[4884]: I1202 01:57:05.032083 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 02 01:57:05 crc kubenswrapper[4884]: I1202 01:57:05.032107 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 02 01:57:05 crc kubenswrapper[4884]: I1202 01:57:05.060118 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-57b6695dcb-8c27t" podStartSLOduration=3.133766826 podStartE2EDuration="6.060097643s" podCreationTimestamp="2025-12-02 01:56:59 +0000 UTC" firstStartedPulling="2025-12-02 01:57:00.591088178 +0000 UTC m=+1237.266925062" lastFinishedPulling="2025-12-02 01:57:03.517418995 +0000 UTC m=+1240.193255879" observedRunningTime="2025-12-02 01:57:05.041400861 +0000 UTC m=+1241.717237755" watchObservedRunningTime="2025-12-02 01:57:05.060097643 +0000 UTC m=+1241.735934547" Dec 02 01:57:05 crc kubenswrapper[4884]: I1202 01:57:05.075162 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7b4498cc99-xx9bb" podStartSLOduration=2.835364268 podStartE2EDuration="6.075143454s" podCreationTimestamp="2025-12-02 01:56:59 +0000 UTC" firstStartedPulling="2025-12-02 01:57:00.278562982 +0000 UTC m=+1236.954399866" lastFinishedPulling="2025-12-02 01:57:03.518342168 +0000 UTC m=+1240.194179052" observedRunningTime="2025-12-02 01:57:05.066239765 +0000 UTC m=+1241.742076669" watchObservedRunningTime="2025-12-02 01:57:05.075143454 +0000 UTC m=+1241.750980338" Dec 02 01:57:07 crc kubenswrapper[4884]: I1202 01:57:07.023830 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 02 01:57:07 crc kubenswrapper[4884]: I1202 01:57:07.027199 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 02 01:57:08 crc kubenswrapper[4884]: I1202 01:57:08.926494 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.075048 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-c9rsj" event={"ID":"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada","Type":"ContainerDied","Data":"18df2fffc1924f7ae65d40273225114fb51661eb6360fa097db4957fef32d835"} Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.075080 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="18df2fffc1924f7ae65d40273225114fb51661eb6360fa097db4957fef32d835" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.075129 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-c9rsj" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.107642 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-db-sync-config-data\") pod \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.107696 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-etc-machine-id\") pod \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.107731 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd49x\" (UniqueName: \"kubernetes.io/projected/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-kube-api-access-wd49x\") pod \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.107789 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-combined-ca-bundle\") pod \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.107845 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f0d68bd4-6c59-4e41-b32e-f7918d3b4ada" (UID: "f0d68bd4-6c59-4e41-b32e-f7918d3b4ada"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.107828 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-config-data\") pod \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.108632 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-scripts\") pod \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\" (UID: \"f0d68bd4-6c59-4e41-b32e-f7918d3b4ada\") " Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.109051 4884 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.115370 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f0d68bd4-6c59-4e41-b32e-f7918d3b4ada" (UID: "f0d68bd4-6c59-4e41-b32e-f7918d3b4ada"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.115625 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-kube-api-access-wd49x" (OuterVolumeSpecName: "kube-api-access-wd49x") pod "f0d68bd4-6c59-4e41-b32e-f7918d3b4ada" (UID: "f0d68bd4-6c59-4e41-b32e-f7918d3b4ada"). InnerVolumeSpecName "kube-api-access-wd49x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.116643 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-scripts" (OuterVolumeSpecName: "scripts") pod "f0d68bd4-6c59-4e41-b32e-f7918d3b4ada" (UID: "f0d68bd4-6c59-4e41-b32e-f7918d3b4ada"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.143936 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0d68bd4-6c59-4e41-b32e-f7918d3b4ada" (UID: "f0d68bd4-6c59-4e41-b32e-f7918d3b4ada"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.177720 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-config-data" (OuterVolumeSpecName: "config-data") pod "f0d68bd4-6c59-4e41-b32e-f7918d3b4ada" (UID: "f0d68bd4-6c59-4e41-b32e-f7918d3b4ada"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.210932 4884 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.210976 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd49x\" (UniqueName: \"kubernetes.io/projected/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-kube-api-access-wd49x\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.210993 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.211004 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.211015 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.337299 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-api-0" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.350348 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-api-0" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.834690 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.895846 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-whpnl"] Dec 02 01:57:09 crc kubenswrapper[4884]: I1202 01:57:09.896352 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55f844cf75-whpnl" podUID="0c25ce5d-09d7-4a8b-be58-baa2d27191e8" containerName="dnsmasq-dns" containerID="cri-o://9e023e5e5819281374724729f6fbf33e9564e4e8e61e4b446f963b387a9c54d8" gracePeriod=10 Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.115127 4884 generic.go:334] "Generic (PLEG): container finished" podID="0c25ce5d-09d7-4a8b-be58-baa2d27191e8" containerID="9e023e5e5819281374724729f6fbf33e9564e4e8e61e4b446f963b387a9c54d8" exitCode=0 Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.115197 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-whpnl" event={"ID":"0c25ce5d-09d7-4a8b-be58-baa2d27191e8","Type":"ContainerDied","Data":"9e023e5e5819281374724729f6fbf33e9564e4e8e61e4b446f963b387a9c54d8"} Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.128255 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45","Type":"ContainerStarted","Data":"5dacf26e11fcbf89cd8ffb628a4bc8ba3af133673e28c4fb688480590471d7d7"} Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.128364 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="ceilometer-central-agent" containerID="cri-o://b40142baf48473a349fd9c65a988c97d8d72dedfe8f3737777935794684cb579" gracePeriod=30 Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.128382 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.128474 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="proxy-httpd" containerID="cri-o://5dacf26e11fcbf89cd8ffb628a4bc8ba3af133673e28c4fb688480590471d7d7" gracePeriod=30 Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.128523 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="sg-core" containerID="cri-o://e5249e12843a0b5a923b2d8cbcb00cd0581382a128766e190f68ffd31688caf4" gracePeriod=30 Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.128889 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="ceilometer-notification-agent" containerID="cri-o://94a6e057cb4483d3a3bf690284e2619434c9de4535aa0e4991c8c869f0dd057e" gracePeriod=30 Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.132780 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5fb984d75d-76xfp" event={"ID":"6059a81d-e008-4cba-8af1-9b093989ddf2","Type":"ContainerStarted","Data":"2a3c14a355fc958c2f2719d83462c906948e928e777a409837c845621a1f85d9"} Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.133144 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.133903 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.140531 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5fb984d75d-76xfp" podUID="6059a81d-e008-4cba-8af1-9b093989ddf2" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.176:9311/healthcheck\": dial tcp 10.217.0.176:9311: connect: connection refused" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.154364 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.078522125 podStartE2EDuration="1m1.154348604s" podCreationTimestamp="2025-12-02 01:56:09 +0000 UTC" firstStartedPulling="2025-12-02 01:56:11.622201079 +0000 UTC m=+1188.298037963" lastFinishedPulling="2025-12-02 01:57:09.698027558 +0000 UTC m=+1246.373864442" observedRunningTime="2025-12-02 01:57:10.15340397 +0000 UTC m=+1246.829240864" watchObservedRunningTime="2025-12-02 01:57:10.154348604 +0000 UTC m=+1246.830185488" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.181602 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5fb984d75d-76xfp" podStartSLOduration=8.181583766 podStartE2EDuration="8.181583766s" podCreationTimestamp="2025-12-02 01:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:57:10.177412643 +0000 UTC m=+1246.853249547" watchObservedRunningTime="2025-12-02 01:57:10.181583766 +0000 UTC m=+1246.857420650" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.202214 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.237325 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 01:57:10 crc kubenswrapper[4884]: E1202 01:57:10.237976 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0d68bd4-6c59-4e41-b32e-f7918d3b4ada" containerName="cinder-db-sync" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.237989 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0d68bd4-6c59-4e41-b32e-f7918d3b4ada" containerName="cinder-db-sync" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.238180 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0d68bd4-6c59-4e41-b32e-f7918d3b4ada" containerName="cinder-db-sync" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.239201 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.242803 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.242957 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-2chcb" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.243003 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.243139 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.263446 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.332673 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.332772 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.332816 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-config-data\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.332855 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.332891 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzt4t\" (UniqueName: \"kubernetes.io/projected/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-kube-api-access-xzt4t\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.332932 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-scripts\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.338272 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-cbbd4"] Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.342303 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.370951 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-cbbd4"] Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.434923 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb4vp\" (UniqueName: \"kubernetes.io/projected/91335ed2-246d-42b7-9d69-7ac7aada058b-kube-api-access-qb4vp\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.435002 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.435038 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-config-data\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.435081 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.435123 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzt4t\" (UniqueName: \"kubernetes.io/projected/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-kube-api-access-xzt4t\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.435143 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-config\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.435182 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-scripts\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.435202 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.435231 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.435268 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.435291 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.435319 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.437000 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.443314 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-scripts\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.456681 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.459233 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.469019 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-config-data\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.472855 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzt4t\" (UniqueName: \"kubernetes.io/projected/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-kube-api-access-xzt4t\") pod \"cinder-scheduler-0\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.538384 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.538482 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-config\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.538534 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.538569 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.538589 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.538618 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb4vp\" (UniqueName: \"kubernetes.io/projected/91335ed2-246d-42b7-9d69-7ac7aada058b-kube-api-access-qb4vp\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.539574 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.540619 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-config\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.541481 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.542155 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.546547 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.567932 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb4vp\" (UniqueName: \"kubernetes.io/projected/91335ed2-246d-42b7-9d69-7ac7aada058b-kube-api-access-qb4vp\") pod \"dnsmasq-dns-5c9776ccc5-cbbd4\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.575649 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.577429 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.580146 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.591737 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.637177 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.678088 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.706224 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.743404 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/384beae3-8e7b-4123-afea-d2fc336cf5ca-etc-machine-id\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.743560 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.743662 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-config-data\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.743775 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-config-data-custom\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.743861 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-scripts\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.743999 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvht2\" (UniqueName: \"kubernetes.io/projected/384beae3-8e7b-4123-afea-d2fc336cf5ca-kube-api-access-zvht2\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.744130 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/384beae3-8e7b-4123-afea-d2fc336cf5ca-logs\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.845940 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-config\") pod \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.845985 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-ovsdbserver-nb\") pod \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.846056 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-ovsdbserver-sb\") pod \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.846131 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-dns-swift-storage-0\") pod \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.846219 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-dns-svc\") pod \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.846244 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbfn7\" (UniqueName: \"kubernetes.io/projected/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-kube-api-access-qbfn7\") pod \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\" (UID: \"0c25ce5d-09d7-4a8b-be58-baa2d27191e8\") " Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.846868 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/384beae3-8e7b-4123-afea-d2fc336cf5ca-logs\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.846948 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/384beae3-8e7b-4123-afea-d2fc336cf5ca-etc-machine-id\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.846974 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.847007 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-config-data\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.847050 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-config-data-custom\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.847070 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-scripts\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.847122 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvht2\" (UniqueName: \"kubernetes.io/projected/384beae3-8e7b-4123-afea-d2fc336cf5ca-kube-api-access-zvht2\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.847670 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/384beae3-8e7b-4123-afea-d2fc336cf5ca-logs\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.849618 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/384beae3-8e7b-4123-afea-d2fc336cf5ca-etc-machine-id\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.887486 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-config-data\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.895044 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-kube-api-access-qbfn7" (OuterVolumeSpecName: "kube-api-access-qbfn7") pod "0c25ce5d-09d7-4a8b-be58-baa2d27191e8" (UID: "0c25ce5d-09d7-4a8b-be58-baa2d27191e8"). InnerVolumeSpecName "kube-api-access-qbfn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.897672 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.897828 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-config-data-custom\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.900389 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvht2\" (UniqueName: \"kubernetes.io/projected/384beae3-8e7b-4123-afea-d2fc336cf5ca-kube-api-access-zvht2\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.903789 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-scripts\") pod \"cinder-api-0\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.938014 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.959595 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbfn7\" (UniqueName: \"kubernetes.io/projected/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-kube-api-access-qbfn7\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.962329 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0c25ce5d-09d7-4a8b-be58-baa2d27191e8" (UID: "0c25ce5d-09d7-4a8b-be58-baa2d27191e8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.962371 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-config" (OuterVolumeSpecName: "config") pod "0c25ce5d-09d7-4a8b-be58-baa2d27191e8" (UID: "0c25ce5d-09d7-4a8b-be58-baa2d27191e8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.995393 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.998816 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0c25ce5d-09d7-4a8b-be58-baa2d27191e8" (UID: "0c25ce5d-09d7-4a8b-be58-baa2d27191e8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:10 crc kubenswrapper[4884]: I1202 01:57:10.998919 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0c25ce5d-09d7-4a8b-be58-baa2d27191e8" (UID: "0c25ce5d-09d7-4a8b-be58-baa2d27191e8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.003935 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0c25ce5d-09d7-4a8b-be58-baa2d27191e8" (UID: "0c25ce5d-09d7-4a8b-be58-baa2d27191e8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.072044 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.072331 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.072340 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.072349 4884 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.072358 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c25ce5d-09d7-4a8b-be58-baa2d27191e8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.157949 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55f844cf75-whpnl" event={"ID":"0c25ce5d-09d7-4a8b-be58-baa2d27191e8","Type":"ContainerDied","Data":"6a769b7aa74d7a6f402a4e164e36bde4ed1dd57ea2a2ff87d9c12a5c2599098b"} Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.157995 4884 scope.go:117] "RemoveContainer" containerID="9e023e5e5819281374724729f6fbf33e9564e4e8e61e4b446f963b387a9c54d8" Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.158132 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55f844cf75-whpnl" Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.174604 4884 generic.go:334] "Generic (PLEG): container finished" podID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerID="5dacf26e11fcbf89cd8ffb628a4bc8ba3af133673e28c4fb688480590471d7d7" exitCode=0 Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.174629 4884 generic.go:334] "Generic (PLEG): container finished" podID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerID="e5249e12843a0b5a923b2d8cbcb00cd0581382a128766e190f68ffd31688caf4" exitCode=2 Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.174637 4884 generic.go:334] "Generic (PLEG): container finished" podID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerID="b40142baf48473a349fd9c65a988c97d8d72dedfe8f3737777935794684cb579" exitCode=0 Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.175512 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45","Type":"ContainerDied","Data":"5dacf26e11fcbf89cd8ffb628a4bc8ba3af133673e28c4fb688480590471d7d7"} Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.175536 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45","Type":"ContainerDied","Data":"e5249e12843a0b5a923b2d8cbcb00cd0581382a128766e190f68ffd31688caf4"} Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.175546 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45","Type":"ContainerDied","Data":"b40142baf48473a349fd9c65a988c97d8d72dedfe8f3737777935794684cb579"} Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.212842 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-whpnl"] Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.214117 4884 scope.go:117] "RemoveContainer" containerID="d1297a53a478a98a781b4e0b8c4cb0ad1f0352a7e94c0e82fb854590cfc3ac66" Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.218234 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55f844cf75-whpnl"] Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.362157 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.524194 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-cbbd4"] Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.624063 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c25ce5d-09d7-4a8b-be58-baa2d27191e8" path="/var/lib/kubelet/pods/0c25ce5d-09d7-4a8b-be58-baa2d27191e8/volumes" Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.637840 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 02 01:57:11 crc kubenswrapper[4884]: I1202 01:57:11.984782 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:57:12 crc kubenswrapper[4884]: I1202 01:57:12.094988 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:57:12 crc kubenswrapper[4884]: I1202 01:57:12.143117 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:57:12 crc kubenswrapper[4884]: I1202 01:57:12.239998 4884 generic.go:334] "Generic (PLEG): container finished" podID="91335ed2-246d-42b7-9d69-7ac7aada058b" containerID="913e9170ef6de8e043f6272ead9c298935c9ab5c5862bd5ca39a2505fd748383" exitCode=0 Dec 02 01:57:12 crc kubenswrapper[4884]: I1202 01:57:12.240087 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" event={"ID":"91335ed2-246d-42b7-9d69-7ac7aada058b","Type":"ContainerDied","Data":"913e9170ef6de8e043f6272ead9c298935c9ab5c5862bd5ca39a2505fd748383"} Dec 02 01:57:12 crc kubenswrapper[4884]: I1202 01:57:12.240112 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" event={"ID":"91335ed2-246d-42b7-9d69-7ac7aada058b","Type":"ContainerStarted","Data":"bb80f41bf6b5d2c09f29fdde80c97df4149f45281103fb9609281a746c663022"} Dec 02 01:57:12 crc kubenswrapper[4884]: I1202 01:57:12.289985 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4","Type":"ContainerStarted","Data":"cd6fc3e4e9929fa45060aa657e328a9079df9aaeb62e1276db94a3a620be1cac"} Dec 02 01:57:12 crc kubenswrapper[4884]: I1202 01:57:12.308948 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"384beae3-8e7b-4123-afea-d2fc336cf5ca","Type":"ContainerStarted","Data":"355f2098b2b3ed9fc2ddabbf4adcb77c5cb785cc8b3021fe76a1783070ae95cd"} Dec 02 01:57:12 crc kubenswrapper[4884]: I1202 01:57:12.611629 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 02 01:57:12 crc kubenswrapper[4884]: I1202 01:57:12.811285 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.134018 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.292168 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7b6477d857-jzgzr" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.316232 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.359616 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" event={"ID":"91335ed2-246d-42b7-9d69-7ac7aada058b","Type":"ContainerStarted","Data":"d27a0f434086c0b10da27c069a025623ffb6e9ac1fe49ff8d19f8ade64719295"} Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.360150 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.410118 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-679d6c649b-jzn47"] Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.410335 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-679d6c649b-jzn47" podUID="321c69df-0c85-4378-af0f-2b127be9b6a9" containerName="neutron-api" containerID="cri-o://c3ab0e53139b21f9409514545ffa66d0e8b941261ba7a639cccb4de8b20b6d66" gracePeriod=30 Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.410613 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-679d6c649b-jzn47" podUID="321c69df-0c85-4378-af0f-2b127be9b6a9" containerName="neutron-httpd" containerID="cri-o://cedf292b3a4697c44d487a2d956d2b7d11751fa4737d94c1814906ca7baaa879" gracePeriod=30 Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.414323 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" podStartSLOduration=3.414309889 podStartE2EDuration="3.414309889s" podCreationTimestamp="2025-12-02 01:57:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:57:13.403905372 +0000 UTC m=+1250.079742266" watchObservedRunningTime="2025-12-02 01:57:13.414309889 +0000 UTC m=+1250.090146783" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.417002 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"384beae3-8e7b-4123-afea-d2fc336cf5ca","Type":"ContainerStarted","Data":"ecaf1c6fbe8b4fcb33fba2de084fe0f9aeae36aef687945f3663c29a85ad4cee"} Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.420065 4884 generic.go:334] "Generic (PLEG): container finished" podID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerID="94a6e057cb4483d3a3bf690284e2619434c9de4535aa0e4991c8c869f0dd057e" exitCode=0 Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.420872 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.421266 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45","Type":"ContainerDied","Data":"94a6e057cb4483d3a3bf690284e2619434c9de4535aa0e4991c8c869f0dd057e"} Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.421292 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45","Type":"ContainerDied","Data":"6d7bdda8240d78eef550bf103054c243a7691255acc6b3f4211e1269eaddd053"} Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.421308 4884 scope.go:117] "RemoveContainer" containerID="5dacf26e11fcbf89cd8ffb628a4bc8ba3af133673e28c4fb688480590471d7d7" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.450395 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-combined-ca-bundle\") pod \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.450471 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-scripts\") pod \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.450510 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-log-httpd\") pod \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.450547 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmt9v\" (UniqueName: \"kubernetes.io/projected/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-kube-api-access-lmt9v\") pod \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.450601 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-run-httpd\") pod \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.450649 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-config-data\") pod \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.450667 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-sg-core-conf-yaml\") pod \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\" (UID: \"8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45\") " Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.453185 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" (UID: "8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.453449 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" (UID: "8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.462539 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-kube-api-access-lmt9v" (OuterVolumeSpecName: "kube-api-access-lmt9v") pod "8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" (UID: "8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45"). InnerVolumeSpecName "kube-api-access-lmt9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.489122 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-scripts" (OuterVolumeSpecName: "scripts") pod "8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" (UID: "8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.552817 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.552839 4884 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.552848 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmt9v\" (UniqueName: \"kubernetes.io/projected/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-kube-api-access-lmt9v\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.552857 4884 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.558900 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" (UID: "8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.600591 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" (UID: "8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.612859 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-config-data" (OuterVolumeSpecName: "config-data") pod "8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" (UID: "8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.654316 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.654340 4884 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.654353 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.764205 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.789567 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.798921 4884 scope.go:117] "RemoveContainer" containerID="e5249e12843a0b5a923b2d8cbcb00cd0581382a128766e190f68ffd31688caf4" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.810924 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:13 crc kubenswrapper[4884]: E1202 01:57:13.811352 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="sg-core" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.811365 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="sg-core" Dec 02 01:57:13 crc kubenswrapper[4884]: E1202 01:57:13.811378 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c25ce5d-09d7-4a8b-be58-baa2d27191e8" containerName="init" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.811387 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c25ce5d-09d7-4a8b-be58-baa2d27191e8" containerName="init" Dec 02 01:57:13 crc kubenswrapper[4884]: E1202 01:57:13.811405 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="ceilometer-notification-agent" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.811411 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="ceilometer-notification-agent" Dec 02 01:57:13 crc kubenswrapper[4884]: E1202 01:57:13.811429 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="ceilometer-central-agent" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.811434 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="ceilometer-central-agent" Dec 02 01:57:13 crc kubenswrapper[4884]: E1202 01:57:13.811443 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="proxy-httpd" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.811450 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="proxy-httpd" Dec 02 01:57:13 crc kubenswrapper[4884]: E1202 01:57:13.811465 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c25ce5d-09d7-4a8b-be58-baa2d27191e8" containerName="dnsmasq-dns" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.811471 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c25ce5d-09d7-4a8b-be58-baa2d27191e8" containerName="dnsmasq-dns" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.811633 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="proxy-httpd" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.811650 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="ceilometer-central-agent" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.811663 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="ceilometer-notification-agent" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.811672 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" containerName="sg-core" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.811687 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c25ce5d-09d7-4a8b-be58-baa2d27191e8" containerName="dnsmasq-dns" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.813493 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.817077 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.821006 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.835670 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.838879 4884 scope.go:117] "RemoveContainer" containerID="94a6e057cb4483d3a3bf690284e2619434c9de4535aa0e4991c8c869f0dd057e" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.891866 4884 scope.go:117] "RemoveContainer" containerID="b40142baf48473a349fd9c65a988c97d8d72dedfe8f3737777935794684cb579" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.963544 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm77p\" (UniqueName: \"kubernetes.io/projected/37b95259-8a7e-43fb-8b52-805528b86103-kube-api-access-qm77p\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.963862 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37b95259-8a7e-43fb-8b52-805528b86103-log-httpd\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.963918 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-config-data\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.963936 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37b95259-8a7e-43fb-8b52-805528b86103-run-httpd\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.963960 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.964006 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.964030 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-scripts\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.987890 4884 scope.go:117] "RemoveContainer" containerID="5dacf26e11fcbf89cd8ffb628a4bc8ba3af133673e28c4fb688480590471d7d7" Dec 02 01:57:13 crc kubenswrapper[4884]: E1202 01:57:13.988892 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dacf26e11fcbf89cd8ffb628a4bc8ba3af133673e28c4fb688480590471d7d7\": container with ID starting with 5dacf26e11fcbf89cd8ffb628a4bc8ba3af133673e28c4fb688480590471d7d7 not found: ID does not exist" containerID="5dacf26e11fcbf89cd8ffb628a4bc8ba3af133673e28c4fb688480590471d7d7" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.988922 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dacf26e11fcbf89cd8ffb628a4bc8ba3af133673e28c4fb688480590471d7d7"} err="failed to get container status \"5dacf26e11fcbf89cd8ffb628a4bc8ba3af133673e28c4fb688480590471d7d7\": rpc error: code = NotFound desc = could not find container \"5dacf26e11fcbf89cd8ffb628a4bc8ba3af133673e28c4fb688480590471d7d7\": container with ID starting with 5dacf26e11fcbf89cd8ffb628a4bc8ba3af133673e28c4fb688480590471d7d7 not found: ID does not exist" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.988942 4884 scope.go:117] "RemoveContainer" containerID="e5249e12843a0b5a923b2d8cbcb00cd0581382a128766e190f68ffd31688caf4" Dec 02 01:57:13 crc kubenswrapper[4884]: E1202 01:57:13.989298 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5249e12843a0b5a923b2d8cbcb00cd0581382a128766e190f68ffd31688caf4\": container with ID starting with e5249e12843a0b5a923b2d8cbcb00cd0581382a128766e190f68ffd31688caf4 not found: ID does not exist" containerID="e5249e12843a0b5a923b2d8cbcb00cd0581382a128766e190f68ffd31688caf4" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.989319 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5249e12843a0b5a923b2d8cbcb00cd0581382a128766e190f68ffd31688caf4"} err="failed to get container status \"e5249e12843a0b5a923b2d8cbcb00cd0581382a128766e190f68ffd31688caf4\": rpc error: code = NotFound desc = could not find container \"e5249e12843a0b5a923b2d8cbcb00cd0581382a128766e190f68ffd31688caf4\": container with ID starting with e5249e12843a0b5a923b2d8cbcb00cd0581382a128766e190f68ffd31688caf4 not found: ID does not exist" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.989334 4884 scope.go:117] "RemoveContainer" containerID="94a6e057cb4483d3a3bf690284e2619434c9de4535aa0e4991c8c869f0dd057e" Dec 02 01:57:13 crc kubenswrapper[4884]: E1202 01:57:13.994991 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94a6e057cb4483d3a3bf690284e2619434c9de4535aa0e4991c8c869f0dd057e\": container with ID starting with 94a6e057cb4483d3a3bf690284e2619434c9de4535aa0e4991c8c869f0dd057e not found: ID does not exist" containerID="94a6e057cb4483d3a3bf690284e2619434c9de4535aa0e4991c8c869f0dd057e" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.995055 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94a6e057cb4483d3a3bf690284e2619434c9de4535aa0e4991c8c869f0dd057e"} err="failed to get container status \"94a6e057cb4483d3a3bf690284e2619434c9de4535aa0e4991c8c869f0dd057e\": rpc error: code = NotFound desc = could not find container \"94a6e057cb4483d3a3bf690284e2619434c9de4535aa0e4991c8c869f0dd057e\": container with ID starting with 94a6e057cb4483d3a3bf690284e2619434c9de4535aa0e4991c8c869f0dd057e not found: ID does not exist" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.995088 4884 scope.go:117] "RemoveContainer" containerID="b40142baf48473a349fd9c65a988c97d8d72dedfe8f3737777935794684cb579" Dec 02 01:57:13 crc kubenswrapper[4884]: E1202 01:57:13.998230 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b40142baf48473a349fd9c65a988c97d8d72dedfe8f3737777935794684cb579\": container with ID starting with b40142baf48473a349fd9c65a988c97d8d72dedfe8f3737777935794684cb579 not found: ID does not exist" containerID="b40142baf48473a349fd9c65a988c97d8d72dedfe8f3737777935794684cb579" Dec 02 01:57:13 crc kubenswrapper[4884]: I1202 01:57:13.998265 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b40142baf48473a349fd9c65a988c97d8d72dedfe8f3737777935794684cb579"} err="failed to get container status \"b40142baf48473a349fd9c65a988c97d8d72dedfe8f3737777935794684cb579\": rpc error: code = NotFound desc = could not find container \"b40142baf48473a349fd9c65a988c97d8d72dedfe8f3737777935794684cb579\": container with ID starting with b40142baf48473a349fd9c65a988c97d8d72dedfe8f3737777935794684cb579 not found: ID does not exist" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.068151 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37b95259-8a7e-43fb-8b52-805528b86103-log-httpd\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.068221 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-config-data\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.068243 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37b95259-8a7e-43fb-8b52-805528b86103-run-httpd\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.068270 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.068314 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.068336 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-scripts\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.068358 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm77p\" (UniqueName: \"kubernetes.io/projected/37b95259-8a7e-43fb-8b52-805528b86103-kube-api-access-qm77p\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.069021 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37b95259-8a7e-43fb-8b52-805528b86103-log-httpd\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.072371 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.072791 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37b95259-8a7e-43fb-8b52-805528b86103-run-httpd\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.096555 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm77p\" (UniqueName: \"kubernetes.io/projected/37b95259-8a7e-43fb-8b52-805528b86103-kube-api-access-qm77p\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.097773 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-config-data\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.103327 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.109643 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-scripts\") pod \"ceilometer-0\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " pod="openstack/ceilometer-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.171952 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.485400 4884 generic.go:334] "Generic (PLEG): container finished" podID="321c69df-0c85-4378-af0f-2b127be9b6a9" containerID="cedf292b3a4697c44d487a2d956d2b7d11751fa4737d94c1814906ca7baaa879" exitCode=0 Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.486480 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-679d6c649b-jzn47" event={"ID":"321c69df-0c85-4378-af0f-2b127be9b6a9","Type":"ContainerDied","Data":"cedf292b3a4697c44d487a2d956d2b7d11751fa4737d94c1814906ca7baaa879"} Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.535284 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4","Type":"ContainerStarted","Data":"d81a727fe5867ce6a12a75b34235e9466ffee30e83084e6cdc2fb553c25b849e"} Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.552123 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="384beae3-8e7b-4123-afea-d2fc336cf5ca" containerName="cinder-api-log" containerID="cri-o://ecaf1c6fbe8b4fcb33fba2de084fe0f9aeae36aef687945f3663c29a85ad4cee" gracePeriod=30 Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.552460 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"384beae3-8e7b-4123-afea-d2fc336cf5ca","Type":"ContainerStarted","Data":"a40dd0c65b503ab2582ab52cea2713dbf77ac35705aed55b3072c898a73a1373"} Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.552502 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.552535 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="384beae3-8e7b-4123-afea-d2fc336cf5ca" containerName="cinder-api" containerID="cri-o://a40dd0c65b503ab2582ab52cea2713dbf77ac35705aed55b3072c898a73a1373" gracePeriod=30 Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.588436 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.588414826 podStartE2EDuration="4.588414826s" podCreationTimestamp="2025-12-02 01:57:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:57:14.577416215 +0000 UTC m=+1251.253253099" watchObservedRunningTime="2025-12-02 01:57:14.588414826 +0000 UTC m=+1251.264251710" Dec 02 01:57:14 crc kubenswrapper[4884]: I1202 01:57:14.894169 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.109262 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-8687f948dd-5zg8m" Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.122416 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.198507 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-55fc6c9b58-9v749"] Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.353905 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-55f844cf75-whpnl" podUID="0c25ce5d-09d7-4a8b-be58-baa2d27191e8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.161:5353: i/o timeout" Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.569288 4884 generic.go:334] "Generic (PLEG): container finished" podID="384beae3-8e7b-4123-afea-d2fc336cf5ca" containerID="a40dd0c65b503ab2582ab52cea2713dbf77ac35705aed55b3072c898a73a1373" exitCode=0 Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.569508 4884 generic.go:334] "Generic (PLEG): container finished" podID="384beae3-8e7b-4123-afea-d2fc336cf5ca" containerID="ecaf1c6fbe8b4fcb33fba2de084fe0f9aeae36aef687945f3663c29a85ad4cee" exitCode=143 Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.569559 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"384beae3-8e7b-4123-afea-d2fc336cf5ca","Type":"ContainerDied","Data":"a40dd0c65b503ab2582ab52cea2713dbf77ac35705aed55b3072c898a73a1373"} Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.569584 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"384beae3-8e7b-4123-afea-d2fc336cf5ca","Type":"ContainerDied","Data":"ecaf1c6fbe8b4fcb33fba2de084fe0f9aeae36aef687945f3663c29a85ad4cee"} Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.593084 4884 generic.go:334] "Generic (PLEG): container finished" podID="321c69df-0c85-4378-af0f-2b127be9b6a9" containerID="c3ab0e53139b21f9409514545ffa66d0e8b941261ba7a639cccb4de8b20b6d66" exitCode=0 Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.593146 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-679d6c649b-jzn47" event={"ID":"321c69df-0c85-4378-af0f-2b127be9b6a9","Type":"ContainerDied","Data":"c3ab0e53139b21f9409514545ffa66d0e8b941261ba7a639cccb4de8b20b6d66"} Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.605829 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37b95259-8a7e-43fb-8b52-805528b86103","Type":"ContainerStarted","Data":"c93a14b311a2a49d068d3cb80b584bd9eca5abd99260c1846bde4523b5f19397"} Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.608533 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-55fc6c9b58-9v749" podUID="59b51ab4-e0c8-42af-92aa-063e66157aa7" containerName="horizon-log" containerID="cri-o://d1f74407d395e0019fa6a1d81737da09d7d7e55fb3d882e664487b5cbff7aa32" gracePeriod=30 Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.609444 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4","Type":"ContainerStarted","Data":"600d4123a3204032be18d4cc42a419e78cb8d87c48eb4a9ce2d3f1b3d1dd6878"} Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.609802 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-55fc6c9b58-9v749" podUID="59b51ab4-e0c8-42af-92aa-063e66157aa7" containerName="horizon" containerID="cri-o://b52f5ae968a87880245e34143fbe5457d552755e5257d2884c65812f7904191b" gracePeriod=30 Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.635294 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45" path="/var/lib/kubelet/pods/8af75a0b-d7d4-4c8f-b2fd-9337b9d59e45/volumes" Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.638052 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.646358 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.38522506 podStartE2EDuration="5.646341576s" podCreationTimestamp="2025-12-02 01:57:10 +0000 UTC" firstStartedPulling="2025-12-02 01:57:11.373316049 +0000 UTC m=+1248.049152923" lastFinishedPulling="2025-12-02 01:57:12.634432555 +0000 UTC m=+1249.310269439" observedRunningTime="2025-12-02 01:57:15.631442318 +0000 UTC m=+1252.307279212" watchObservedRunningTime="2025-12-02 01:57:15.646341576 +0000 UTC m=+1252.322178460" Dec 02 01:57:15 crc kubenswrapper[4884]: I1202 01:57:15.874367 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.013359 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/384beae3-8e7b-4123-afea-d2fc336cf5ca-etc-machine-id\") pod \"384beae3-8e7b-4123-afea-d2fc336cf5ca\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.013716 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-config-data\") pod \"384beae3-8e7b-4123-afea-d2fc336cf5ca\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.013665 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/384beae3-8e7b-4123-afea-d2fc336cf5ca-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "384beae3-8e7b-4123-afea-d2fc336cf5ca" (UID: "384beae3-8e7b-4123-afea-d2fc336cf5ca"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.013859 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-config-data-custom\") pod \"384beae3-8e7b-4123-afea-d2fc336cf5ca\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.013889 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-combined-ca-bundle\") pod \"384beae3-8e7b-4123-afea-d2fc336cf5ca\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.014762 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvht2\" (UniqueName: \"kubernetes.io/projected/384beae3-8e7b-4123-afea-d2fc336cf5ca-kube-api-access-zvht2\") pod \"384beae3-8e7b-4123-afea-d2fc336cf5ca\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.014839 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-scripts\") pod \"384beae3-8e7b-4123-afea-d2fc336cf5ca\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.014972 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/384beae3-8e7b-4123-afea-d2fc336cf5ca-logs\") pod \"384beae3-8e7b-4123-afea-d2fc336cf5ca\" (UID: \"384beae3-8e7b-4123-afea-d2fc336cf5ca\") " Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.015380 4884 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/384beae3-8e7b-4123-afea-d2fc336cf5ca-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.015982 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/384beae3-8e7b-4123-afea-d2fc336cf5ca-logs" (OuterVolumeSpecName: "logs") pod "384beae3-8e7b-4123-afea-d2fc336cf5ca" (UID: "384beae3-8e7b-4123-afea-d2fc336cf5ca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.019175 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "384beae3-8e7b-4123-afea-d2fc336cf5ca" (UID: "384beae3-8e7b-4123-afea-d2fc336cf5ca"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.027890 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-scripts" (OuterVolumeSpecName: "scripts") pod "384beae3-8e7b-4123-afea-d2fc336cf5ca" (UID: "384beae3-8e7b-4123-afea-d2fc336cf5ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.027928 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/384beae3-8e7b-4123-afea-d2fc336cf5ca-kube-api-access-zvht2" (OuterVolumeSpecName: "kube-api-access-zvht2") pod "384beae3-8e7b-4123-afea-d2fc336cf5ca" (UID: "384beae3-8e7b-4123-afea-d2fc336cf5ca"). InnerVolumeSpecName "kube-api-access-zvht2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.062912 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "384beae3-8e7b-4123-afea-d2fc336cf5ca" (UID: "384beae3-8e7b-4123-afea-d2fc336cf5ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.076899 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-config-data" (OuterVolumeSpecName: "config-data") pod "384beae3-8e7b-4123-afea-d2fc336cf5ca" (UID: "384beae3-8e7b-4123-afea-d2fc336cf5ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.116704 4884 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.116732 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.116753 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvht2\" (UniqueName: \"kubernetes.io/projected/384beae3-8e7b-4123-afea-d2fc336cf5ca-kube-api-access-zvht2\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.116764 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.116772 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/384beae3-8e7b-4123-afea-d2fc336cf5ca-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.116780 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/384beae3-8e7b-4123-afea-d2fc336cf5ca-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.159758 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.217555 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-ovndb-tls-certs\") pod \"321c69df-0c85-4378-af0f-2b127be9b6a9\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.217638 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-httpd-config\") pod \"321c69df-0c85-4378-af0f-2b127be9b6a9\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.217707 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-combined-ca-bundle\") pod \"321c69df-0c85-4378-af0f-2b127be9b6a9\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.217769 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmzzz\" (UniqueName: \"kubernetes.io/projected/321c69df-0c85-4378-af0f-2b127be9b6a9-kube-api-access-bmzzz\") pod \"321c69df-0c85-4378-af0f-2b127be9b6a9\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.217838 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-config\") pod \"321c69df-0c85-4378-af0f-2b127be9b6a9\" (UID: \"321c69df-0c85-4378-af0f-2b127be9b6a9\") " Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.221880 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/321c69df-0c85-4378-af0f-2b127be9b6a9-kube-api-access-bmzzz" (OuterVolumeSpecName: "kube-api-access-bmzzz") pod "321c69df-0c85-4378-af0f-2b127be9b6a9" (UID: "321c69df-0c85-4378-af0f-2b127be9b6a9"). InnerVolumeSpecName "kube-api-access-bmzzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.225017 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "321c69df-0c85-4378-af0f-2b127be9b6a9" (UID: "321c69df-0c85-4378-af0f-2b127be9b6a9"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.300023 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-config" (OuterVolumeSpecName: "config") pod "321c69df-0c85-4378-af0f-2b127be9b6a9" (UID: "321c69df-0c85-4378-af0f-2b127be9b6a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.311401 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "321c69df-0c85-4378-af0f-2b127be9b6a9" (UID: "321c69df-0c85-4378-af0f-2b127be9b6a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.320289 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.320322 4884 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.320336 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.320350 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmzzz\" (UniqueName: \"kubernetes.io/projected/321c69df-0c85-4378-af0f-2b127be9b6a9-kube-api-access-bmzzz\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.363846 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "321c69df-0c85-4378-af0f-2b127be9b6a9" (UID: "321c69df-0c85-4378-af0f-2b127be9b6a9"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.422255 4884 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/321c69df-0c85-4378-af0f-2b127be9b6a9-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.622051 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-679d6c649b-jzn47" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.622034 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-679d6c649b-jzn47" event={"ID":"321c69df-0c85-4378-af0f-2b127be9b6a9","Type":"ContainerDied","Data":"1d2295a94cc5d95977813bafce53b7133a267323e5b44d14a37fe3bcdfa4eb4e"} Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.622427 4884 scope.go:117] "RemoveContainer" containerID="cedf292b3a4697c44d487a2d956d2b7d11751fa4737d94c1814906ca7baaa879" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.624670 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37b95259-8a7e-43fb-8b52-805528b86103","Type":"ContainerStarted","Data":"cde32bfe32f8f53ffb99fa8611f3a610e84f6515870159f5cc71820c3071676b"} Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.624712 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37b95259-8a7e-43fb-8b52-805528b86103","Type":"ContainerStarted","Data":"3722720b386461ac68ae9a1937a67f53f525d12995f309ceccc98bcd4152dbfa"} Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.632403 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.632547 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"384beae3-8e7b-4123-afea-d2fc336cf5ca","Type":"ContainerDied","Data":"355f2098b2b3ed9fc2ddabbf4adcb77c5cb785cc8b3021fe76a1783070ae95cd"} Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.666765 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-679d6c649b-jzn47"] Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.672736 4884 scope.go:117] "RemoveContainer" containerID="c3ab0e53139b21f9409514545ffa66d0e8b941261ba7a639cccb4de8b20b6d66" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.677081 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-679d6c649b-jzn47"] Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.690109 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.701455 4884 scope.go:117] "RemoveContainer" containerID="a40dd0c65b503ab2582ab52cea2713dbf77ac35705aed55b3072c898a73a1373" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.702960 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.723134 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 02 01:57:16 crc kubenswrapper[4884]: E1202 01:57:16.723499 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="384beae3-8e7b-4123-afea-d2fc336cf5ca" containerName="cinder-api" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.723509 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="384beae3-8e7b-4123-afea-d2fc336cf5ca" containerName="cinder-api" Dec 02 01:57:16 crc kubenswrapper[4884]: E1202 01:57:16.723518 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="384beae3-8e7b-4123-afea-d2fc336cf5ca" containerName="cinder-api-log" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.723524 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="384beae3-8e7b-4123-afea-d2fc336cf5ca" containerName="cinder-api-log" Dec 02 01:57:16 crc kubenswrapper[4884]: E1202 01:57:16.723547 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="321c69df-0c85-4378-af0f-2b127be9b6a9" containerName="neutron-api" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.723552 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="321c69df-0c85-4378-af0f-2b127be9b6a9" containerName="neutron-api" Dec 02 01:57:16 crc kubenswrapper[4884]: E1202 01:57:16.723562 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="321c69df-0c85-4378-af0f-2b127be9b6a9" containerName="neutron-httpd" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.723567 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="321c69df-0c85-4378-af0f-2b127be9b6a9" containerName="neutron-httpd" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.723774 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="321c69df-0c85-4378-af0f-2b127be9b6a9" containerName="neutron-httpd" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.723789 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="384beae3-8e7b-4123-afea-d2fc336cf5ca" containerName="cinder-api-log" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.723798 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="384beae3-8e7b-4123-afea-d2fc336cf5ca" containerName="cinder-api" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.723811 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="321c69df-0c85-4378-af0f-2b127be9b6a9" containerName="neutron-api" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.724653 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.738359 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.762780 4884 scope.go:117] "RemoveContainer" containerID="ecaf1c6fbe8b4fcb33fba2de084fe0f9aeae36aef687945f3663c29a85ad4cee" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.762990 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.763091 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.763273 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.832107 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.832150 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.832185 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e174ee23-a3a9-4d96-a76f-d63e984698d2-logs\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.832207 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e174ee23-a3a9-4d96-a76f-d63e984698d2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.832258 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc55v\" (UniqueName: \"kubernetes.io/projected/e174ee23-a3a9-4d96-a76f-d63e984698d2-kube-api-access-gc55v\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.832290 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.832310 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-scripts\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.832338 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-config-data-custom\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.832364 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-config-data\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.934035 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.934074 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-scripts\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.934106 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-config-data-custom\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.934130 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-config-data\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.934203 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.934220 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.934247 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e174ee23-a3a9-4d96-a76f-d63e984698d2-logs\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.934262 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e174ee23-a3a9-4d96-a76f-d63e984698d2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.934310 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc55v\" (UniqueName: \"kubernetes.io/projected/e174ee23-a3a9-4d96-a76f-d63e984698d2-kube-api-access-gc55v\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.934950 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e174ee23-a3a9-4d96-a76f-d63e984698d2-logs\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.934998 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e174ee23-a3a9-4d96-a76f-d63e984698d2-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.938638 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.941883 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-config-data-custom\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.943025 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.943447 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-scripts\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.943853 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-config-data\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.946865 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e174ee23-a3a9-4d96-a76f-d63e984698d2-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.950677 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc55v\" (UniqueName: \"kubernetes.io/projected/e174ee23-a3a9-4d96-a76f-d63e984698d2-kube-api-access-gc55v\") pod \"cinder-api-0\" (UID: \"e174ee23-a3a9-4d96-a76f-d63e984698d2\") " pod="openstack/cinder-api-0" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.971882 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.971937 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.971984 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.972706 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"58124d3cd233d6efb787d810e5402863819796c49f3516cfa49fd6248e566581"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 01:57:16 crc kubenswrapper[4884]: I1202 01:57:16.972780 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://58124d3cd233d6efb787d810e5402863819796c49f3516cfa49fd6248e566581" gracePeriod=600 Dec 02 01:57:17 crc kubenswrapper[4884]: I1202 01:57:17.077008 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 02 01:57:17 crc kubenswrapper[4884]: I1202 01:57:17.592976 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 02 01:57:17 crc kubenswrapper[4884]: W1202 01:57:17.608706 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode174ee23_a3a9_4d96_a76f_d63e984698d2.slice/crio-cead5c84af1ce8a45c9ab902589e5cbf4fc66da3a328eb58538447f2850ab38a WatchSource:0}: Error finding container cead5c84af1ce8a45c9ab902589e5cbf4fc66da3a328eb58538447f2850ab38a: Status 404 returned error can't find the container with id cead5c84af1ce8a45c9ab902589e5cbf4fc66da3a328eb58538447f2850ab38a Dec 02 01:57:17 crc kubenswrapper[4884]: I1202 01:57:17.624971 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="321c69df-0c85-4378-af0f-2b127be9b6a9" path="/var/lib/kubelet/pods/321c69df-0c85-4378-af0f-2b127be9b6a9/volumes" Dec 02 01:57:17 crc kubenswrapper[4884]: I1202 01:57:17.625731 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="384beae3-8e7b-4123-afea-d2fc336cf5ca" path="/var/lib/kubelet/pods/384beae3-8e7b-4123-afea-d2fc336cf5ca/volumes" Dec 02 01:57:17 crc kubenswrapper[4884]: I1202 01:57:17.640313 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e174ee23-a3a9-4d96-a76f-d63e984698d2","Type":"ContainerStarted","Data":"cead5c84af1ce8a45c9ab902589e5cbf4fc66da3a328eb58538447f2850ab38a"} Dec 02 01:57:17 crc kubenswrapper[4884]: I1202 01:57:17.644698 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37b95259-8a7e-43fb-8b52-805528b86103","Type":"ContainerStarted","Data":"6bf6a42282d1b7843d02c05bc9da35e24e117c81e0cbc65d9241aed1334871c0"} Dec 02 01:57:17 crc kubenswrapper[4884]: I1202 01:57:17.651569 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="58124d3cd233d6efb787d810e5402863819796c49f3516cfa49fd6248e566581" exitCode=0 Dec 02 01:57:17 crc kubenswrapper[4884]: I1202 01:57:17.651650 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"58124d3cd233d6efb787d810e5402863819796c49f3516cfa49fd6248e566581"} Dec 02 01:57:17 crc kubenswrapper[4884]: I1202 01:57:17.651727 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"5e633af905e3298525fa039bc54a435b8ffe0999ec52daa8e4c43c66ebc12f1d"} Dec 02 01:57:17 crc kubenswrapper[4884]: I1202 01:57:17.651786 4884 scope.go:117] "RemoveContainer" containerID="7b403650e63fc1ca4927612983b11a4a56c5610f07757026da030e5a5f4be8dc" Dec 02 01:57:18 crc kubenswrapper[4884]: I1202 01:57:18.699957 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e174ee23-a3a9-4d96-a76f-d63e984698d2","Type":"ContainerStarted","Data":"08e13e5eecf9b199476c646747e0587d9163297ed9f86bf4133bcc9698b2aebd"} Dec 02 01:57:18 crc kubenswrapper[4884]: I1202 01:57:18.887116 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-595d8586d8-j5926" Dec 02 01:57:19 crc kubenswrapper[4884]: I1202 01:57:19.136506 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-55fc6c9b58-9v749" podUID="59b51ab4-e0c8-42af-92aa-063e66157aa7" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.159:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.159:8443: connect: connection refused" Dec 02 01:57:19 crc kubenswrapper[4884]: I1202 01:57:19.446330 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-595d8586d8-j5926" Dec 02 01:57:19 crc kubenswrapper[4884]: I1202 01:57:19.711186 4884 generic.go:334] "Generic (PLEG): container finished" podID="59b51ab4-e0c8-42af-92aa-063e66157aa7" containerID="b52f5ae968a87880245e34143fbe5457d552755e5257d2884c65812f7904191b" exitCode=0 Dec 02 01:57:19 crc kubenswrapper[4884]: I1202 01:57:19.711257 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55fc6c9b58-9v749" event={"ID":"59b51ab4-e0c8-42af-92aa-063e66157aa7","Type":"ContainerDied","Data":"b52f5ae968a87880245e34143fbe5457d552755e5257d2884c65812f7904191b"} Dec 02 01:57:19 crc kubenswrapper[4884]: I1202 01:57:19.714243 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37b95259-8a7e-43fb-8b52-805528b86103","Type":"ContainerStarted","Data":"744158f31b0164b02b4dd08b62bcd62cda42580407a2a9bed1fd25697ab0e6b1"} Dec 02 01:57:19 crc kubenswrapper[4884]: I1202 01:57:19.715618 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 01:57:19 crc kubenswrapper[4884]: I1202 01:57:19.719454 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e174ee23-a3a9-4d96-a76f-d63e984698d2","Type":"ContainerStarted","Data":"d47db25bdffda1080d6a096658e6e721b9cf3f1878798ada6280de2de56e77ba"} Dec 02 01:57:19 crc kubenswrapper[4884]: I1202 01:57:19.719570 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 02 01:57:19 crc kubenswrapper[4884]: I1202 01:57:19.747171 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.818170588 podStartE2EDuration="6.74714576s" podCreationTimestamp="2025-12-02 01:57:13 +0000 UTC" firstStartedPulling="2025-12-02 01:57:14.914855396 +0000 UTC m=+1251.590692280" lastFinishedPulling="2025-12-02 01:57:18.843830568 +0000 UTC m=+1255.519667452" observedRunningTime="2025-12-02 01:57:19.741605383 +0000 UTC m=+1256.417442267" watchObservedRunningTime="2025-12-02 01:57:19.74714576 +0000 UTC m=+1256.422982664" Dec 02 01:57:19 crc kubenswrapper[4884]: I1202 01:57:19.768182 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.768165629 podStartE2EDuration="3.768165629s" podCreationTimestamp="2025-12-02 01:57:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:57:19.760789917 +0000 UTC m=+1256.436626811" watchObservedRunningTime="2025-12-02 01:57:19.768165629 +0000 UTC m=+1256.444002513" Dec 02 01:57:19 crc kubenswrapper[4884]: I1202 01:57:19.856780 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5fb984d75d-76xfp" Dec 02 01:57:19 crc kubenswrapper[4884]: I1202 01:57:19.906939 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6588b54d58-b7m95"] Dec 02 01:57:19 crc kubenswrapper[4884]: I1202 01:57:19.907199 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6588b54d58-b7m95" podUID="e4536243-ac9e-4f38-ac2f-a8649e3b4b22" containerName="barbican-api-log" containerID="cri-o://208eb2afa2b522caa1da161847bde9c2d7cee195285c683fa02623a06b368fdc" gracePeriod=30 Dec 02 01:57:19 crc kubenswrapper[4884]: I1202 01:57:19.907285 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6588b54d58-b7m95" podUID="e4536243-ac9e-4f38-ac2f-a8649e3b4b22" containerName="barbican-api" containerID="cri-o://2182bc11bb7a544dd3b8deee688205b5ef124cc4136a5fc9e6c6ee486e7e245f" gracePeriod=30 Dec 02 01:57:20 crc kubenswrapper[4884]: I1202 01:57:20.708025 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:57:20 crc kubenswrapper[4884]: I1202 01:57:20.766726 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-2j759"] Dec 02 01:57:20 crc kubenswrapper[4884]: I1202 01:57:20.767019 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-2j759" podUID="9a4e2ac4-dc7b-46f4-bee0-7e6670b07990" containerName="dnsmasq-dns" containerID="cri-o://74dfaa214b6c87de35bf84429f2cc26ce9f16580512e52fc422389f7f16e4e06" gracePeriod=10 Dec 02 01:57:20 crc kubenswrapper[4884]: I1202 01:57:20.794016 4884 generic.go:334] "Generic (PLEG): container finished" podID="e4536243-ac9e-4f38-ac2f-a8649e3b4b22" containerID="208eb2afa2b522caa1da161847bde9c2d7cee195285c683fa02623a06b368fdc" exitCode=143 Dec 02 01:57:20 crc kubenswrapper[4884]: I1202 01:57:20.794302 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6588b54d58-b7m95" event={"ID":"e4536243-ac9e-4f38-ac2f-a8649e3b4b22","Type":"ContainerDied","Data":"208eb2afa2b522caa1da161847bde9c2d7cee195285c683fa02623a06b368fdc"} Dec 02 01:57:20 crc kubenswrapper[4884]: I1202 01:57:20.974146 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.042274 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.349062 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.444650 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-ovsdbserver-nb\") pod \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.444765 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-dns-svc\") pod \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.444909 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgfj8\" (UniqueName: \"kubernetes.io/projected/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-kube-api-access-fgfj8\") pod \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.445039 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-ovsdbserver-sb\") pod \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.445124 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-dns-swift-storage-0\") pod \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.446290 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-config\") pod \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\" (UID: \"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990\") " Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.450628 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-kube-api-access-fgfj8" (OuterVolumeSpecName: "kube-api-access-fgfj8") pod "9a4e2ac4-dc7b-46f4-bee0-7e6670b07990" (UID: "9a4e2ac4-dc7b-46f4-bee0-7e6670b07990"). InnerVolumeSpecName "kube-api-access-fgfj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.500425 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9a4e2ac4-dc7b-46f4-bee0-7e6670b07990" (UID: "9a4e2ac4-dc7b-46f4-bee0-7e6670b07990"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.513651 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9a4e2ac4-dc7b-46f4-bee0-7e6670b07990" (UID: "9a4e2ac4-dc7b-46f4-bee0-7e6670b07990"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.519354 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9a4e2ac4-dc7b-46f4-bee0-7e6670b07990" (UID: "9a4e2ac4-dc7b-46f4-bee0-7e6670b07990"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.530269 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9a4e2ac4-dc7b-46f4-bee0-7e6670b07990" (UID: "9a4e2ac4-dc7b-46f4-bee0-7e6670b07990"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.531886 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-config" (OuterVolumeSpecName: "config") pod "9a4e2ac4-dc7b-46f4-bee0-7e6670b07990" (UID: "9a4e2ac4-dc7b-46f4-bee0-7e6670b07990"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.549359 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.549397 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.549408 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgfj8\" (UniqueName: \"kubernetes.io/projected/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-kube-api-access-fgfj8\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.549422 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.549434 4884 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.549447 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.812628 4884 generic.go:334] "Generic (PLEG): container finished" podID="9a4e2ac4-dc7b-46f4-bee0-7e6670b07990" containerID="74dfaa214b6c87de35bf84429f2cc26ce9f16580512e52fc422389f7f16e4e06" exitCode=0 Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.813762 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-2j759" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.814186 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-2j759" event={"ID":"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990","Type":"ContainerDied","Data":"74dfaa214b6c87de35bf84429f2cc26ce9f16580512e52fc422389f7f16e4e06"} Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.814213 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-2j759" event={"ID":"9a4e2ac4-dc7b-46f4-bee0-7e6670b07990","Type":"ContainerDied","Data":"00170689d0d5774d471c147bf0e58ff1e9ac8799f0a93b4f73ebffb783b102de"} Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.814229 4884 scope.go:117] "RemoveContainer" containerID="74dfaa214b6c87de35bf84429f2cc26ce9f16580512e52fc422389f7f16e4e06" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.814398 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" containerName="cinder-scheduler" containerID="cri-o://d81a727fe5867ce6a12a75b34235e9466ffee30e83084e6cdc2fb553c25b849e" gracePeriod=30 Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.814934 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" containerName="probe" containerID="cri-o://600d4123a3204032be18d4cc42a419e78cb8d87c48eb4a9ce2d3f1b3d1dd6878" gracePeriod=30 Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.872780 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-2j759"] Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.881037 4884 scope.go:117] "RemoveContainer" containerID="3cf5e61393f4ec424c6be9f60d610c754eaf902dfa8813b3442375974ac4945f" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.895227 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-2j759"] Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.939639 4884 scope.go:117] "RemoveContainer" containerID="74dfaa214b6c87de35bf84429f2cc26ce9f16580512e52fc422389f7f16e4e06" Dec 02 01:57:21 crc kubenswrapper[4884]: E1202 01:57:21.940096 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74dfaa214b6c87de35bf84429f2cc26ce9f16580512e52fc422389f7f16e4e06\": container with ID starting with 74dfaa214b6c87de35bf84429f2cc26ce9f16580512e52fc422389f7f16e4e06 not found: ID does not exist" containerID="74dfaa214b6c87de35bf84429f2cc26ce9f16580512e52fc422389f7f16e4e06" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.940120 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74dfaa214b6c87de35bf84429f2cc26ce9f16580512e52fc422389f7f16e4e06"} err="failed to get container status \"74dfaa214b6c87de35bf84429f2cc26ce9f16580512e52fc422389f7f16e4e06\": rpc error: code = NotFound desc = could not find container \"74dfaa214b6c87de35bf84429f2cc26ce9f16580512e52fc422389f7f16e4e06\": container with ID starting with 74dfaa214b6c87de35bf84429f2cc26ce9f16580512e52fc422389f7f16e4e06 not found: ID does not exist" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.940146 4884 scope.go:117] "RemoveContainer" containerID="3cf5e61393f4ec424c6be9f60d610c754eaf902dfa8813b3442375974ac4945f" Dec 02 01:57:21 crc kubenswrapper[4884]: E1202 01:57:21.940325 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cf5e61393f4ec424c6be9f60d610c754eaf902dfa8813b3442375974ac4945f\": container with ID starting with 3cf5e61393f4ec424c6be9f60d610c754eaf902dfa8813b3442375974ac4945f not found: ID does not exist" containerID="3cf5e61393f4ec424c6be9f60d610c754eaf902dfa8813b3442375974ac4945f" Dec 02 01:57:21 crc kubenswrapper[4884]: I1202 01:57:21.940342 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cf5e61393f4ec424c6be9f60d610c754eaf902dfa8813b3442375974ac4945f"} err="failed to get container status \"3cf5e61393f4ec424c6be9f60d610c754eaf902dfa8813b3442375974ac4945f\": rpc error: code = NotFound desc = could not find container \"3cf5e61393f4ec424c6be9f60d610c754eaf902dfa8813b3442375974ac4945f\": container with ID starting with 3cf5e61393f4ec424c6be9f60d610c754eaf902dfa8813b3442375974ac4945f not found: ID does not exist" Dec 02 01:57:22 crc kubenswrapper[4884]: I1202 01:57:22.825252 4884 generic.go:334] "Generic (PLEG): container finished" podID="df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" containerID="600d4123a3204032be18d4cc42a419e78cb8d87c48eb4a9ce2d3f1b3d1dd6878" exitCode=0 Dec 02 01:57:22 crc kubenswrapper[4884]: I1202 01:57:22.826561 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4","Type":"ContainerDied","Data":"600d4123a3204032be18d4cc42a419e78cb8d87c48eb4a9ce2d3f1b3d1dd6878"} Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.380989 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-87844dc44-b7fcj" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.542291 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.624266 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a4e2ac4-dc7b-46f4-bee0-7e6670b07990" path="/var/lib/kubelet/pods/9a4e2ac4-dc7b-46f4-bee0-7e6670b07990/volumes" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.693025 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-config-data\") pod \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.693140 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-combined-ca-bundle\") pod \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.693179 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-logs\") pod \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.693232 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpkz8\" (UniqueName: \"kubernetes.io/projected/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-kube-api-access-hpkz8\") pod \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.693337 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-config-data-custom\") pod \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\" (UID: \"e4536243-ac9e-4f38-ac2f-a8649e3b4b22\") " Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.695028 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-logs" (OuterVolumeSpecName: "logs") pod "e4536243-ac9e-4f38-ac2f-a8649e3b4b22" (UID: "e4536243-ac9e-4f38-ac2f-a8649e3b4b22"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.699867 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e4536243-ac9e-4f38-ac2f-a8649e3b4b22" (UID: "e4536243-ac9e-4f38-ac2f-a8649e3b4b22"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.702062 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-kube-api-access-hpkz8" (OuterVolumeSpecName: "kube-api-access-hpkz8") pod "e4536243-ac9e-4f38-ac2f-a8649e3b4b22" (UID: "e4536243-ac9e-4f38-ac2f-a8649e3b4b22"). InnerVolumeSpecName "kube-api-access-hpkz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.724215 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4536243-ac9e-4f38-ac2f-a8649e3b4b22" (UID: "e4536243-ac9e-4f38-ac2f-a8649e3b4b22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.759973 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-config-data" (OuterVolumeSpecName: "config-data") pod "e4536243-ac9e-4f38-ac2f-a8649e3b4b22" (UID: "e4536243-ac9e-4f38-ac2f-a8649e3b4b22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.800531 4884 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.800565 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.800575 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.800584 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.800593 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpkz8\" (UniqueName: \"kubernetes.io/projected/e4536243-ac9e-4f38-ac2f-a8649e3b4b22-kube-api-access-hpkz8\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.841230 4884 generic.go:334] "Generic (PLEG): container finished" podID="e4536243-ac9e-4f38-ac2f-a8649e3b4b22" containerID="2182bc11bb7a544dd3b8deee688205b5ef124cc4136a5fc9e6c6ee486e7e245f" exitCode=0 Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.841270 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6588b54d58-b7m95" event={"ID":"e4536243-ac9e-4f38-ac2f-a8649e3b4b22","Type":"ContainerDied","Data":"2182bc11bb7a544dd3b8deee688205b5ef124cc4136a5fc9e6c6ee486e7e245f"} Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.841305 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6588b54d58-b7m95" event={"ID":"e4536243-ac9e-4f38-ac2f-a8649e3b4b22","Type":"ContainerDied","Data":"97cad8e2c94ff5bed01a28844faa38aae1e69a1c88b142d2a0b8bf16ff38ac84"} Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.841323 4884 scope.go:117] "RemoveContainer" containerID="2182bc11bb7a544dd3b8deee688205b5ef124cc4136a5fc9e6c6ee486e7e245f" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.841346 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6588b54d58-b7m95" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.881444 4884 scope.go:117] "RemoveContainer" containerID="208eb2afa2b522caa1da161847bde9c2d7cee195285c683fa02623a06b368fdc" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.892066 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6588b54d58-b7m95"] Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.900604 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6588b54d58-b7m95"] Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.905389 4884 scope.go:117] "RemoveContainer" containerID="2182bc11bb7a544dd3b8deee688205b5ef124cc4136a5fc9e6c6ee486e7e245f" Dec 02 01:57:23 crc kubenswrapper[4884]: E1202 01:57:23.905921 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2182bc11bb7a544dd3b8deee688205b5ef124cc4136a5fc9e6c6ee486e7e245f\": container with ID starting with 2182bc11bb7a544dd3b8deee688205b5ef124cc4136a5fc9e6c6ee486e7e245f not found: ID does not exist" containerID="2182bc11bb7a544dd3b8deee688205b5ef124cc4136a5fc9e6c6ee486e7e245f" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.905959 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2182bc11bb7a544dd3b8deee688205b5ef124cc4136a5fc9e6c6ee486e7e245f"} err="failed to get container status \"2182bc11bb7a544dd3b8deee688205b5ef124cc4136a5fc9e6c6ee486e7e245f\": rpc error: code = NotFound desc = could not find container \"2182bc11bb7a544dd3b8deee688205b5ef124cc4136a5fc9e6c6ee486e7e245f\": container with ID starting with 2182bc11bb7a544dd3b8deee688205b5ef124cc4136a5fc9e6c6ee486e7e245f not found: ID does not exist" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.905987 4884 scope.go:117] "RemoveContainer" containerID="208eb2afa2b522caa1da161847bde9c2d7cee195285c683fa02623a06b368fdc" Dec 02 01:57:23 crc kubenswrapper[4884]: E1202 01:57:23.906319 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"208eb2afa2b522caa1da161847bde9c2d7cee195285c683fa02623a06b368fdc\": container with ID starting with 208eb2afa2b522caa1da161847bde9c2d7cee195285c683fa02623a06b368fdc not found: ID does not exist" containerID="208eb2afa2b522caa1da161847bde9c2d7cee195285c683fa02623a06b368fdc" Dec 02 01:57:23 crc kubenswrapper[4884]: I1202 01:57:23.906343 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"208eb2afa2b522caa1da161847bde9c2d7cee195285c683fa02623a06b368fdc"} err="failed to get container status \"208eb2afa2b522caa1da161847bde9c2d7cee195285c683fa02623a06b368fdc\": rpc error: code = NotFound desc = could not find container \"208eb2afa2b522caa1da161847bde9c2d7cee195285c683fa02623a06b368fdc\": container with ID starting with 208eb2afa2b522caa1da161847bde9c2d7cee195285c683fa02623a06b368fdc not found: ID does not exist" Dec 02 01:57:25 crc kubenswrapper[4884]: I1202 01:57:25.624483 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4536243-ac9e-4f38-ac2f-a8649e3b4b22" path="/var/lib/kubelet/pods/e4536243-ac9e-4f38-ac2f-a8649e3b4b22/volumes" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.037965 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 02 01:57:26 crc kubenswrapper[4884]: E1202 01:57:26.038721 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4e2ac4-dc7b-46f4-bee0-7e6670b07990" containerName="init" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.038802 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4e2ac4-dc7b-46f4-bee0-7e6670b07990" containerName="init" Dec 02 01:57:26 crc kubenswrapper[4884]: E1202 01:57:26.038845 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4536243-ac9e-4f38-ac2f-a8649e3b4b22" containerName="barbican-api-log" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.038860 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4536243-ac9e-4f38-ac2f-a8649e3b4b22" containerName="barbican-api-log" Dec 02 01:57:26 crc kubenswrapper[4884]: E1202 01:57:26.038906 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4e2ac4-dc7b-46f4-bee0-7e6670b07990" containerName="dnsmasq-dns" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.038923 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4e2ac4-dc7b-46f4-bee0-7e6670b07990" containerName="dnsmasq-dns" Dec 02 01:57:26 crc kubenswrapper[4884]: E1202 01:57:26.038945 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4536243-ac9e-4f38-ac2f-a8649e3b4b22" containerName="barbican-api" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.038961 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4536243-ac9e-4f38-ac2f-a8649e3b4b22" containerName="barbican-api" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.039385 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4e2ac4-dc7b-46f4-bee0-7e6670b07990" containerName="dnsmasq-dns" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.039435 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4536243-ac9e-4f38-ac2f-a8649e3b4b22" containerName="barbican-api-log" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.039497 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4536243-ac9e-4f38-ac2f-a8649e3b4b22" containerName="barbican-api" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.042180 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.045708 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-psbdr" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.045877 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.045945 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.052219 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.144105 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6fa5609-65ed-4b78-a850-4be3eb166ec9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.144183 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b6fa5609-65ed-4b78-a850-4be3eb166ec9-openstack-config\") pod \"openstackclient\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.144247 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b6fa5609-65ed-4b78-a850-4be3eb166ec9-openstack-config-secret\") pod \"openstackclient\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.144321 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jjgr\" (UniqueName: \"kubernetes.io/projected/b6fa5609-65ed-4b78-a850-4be3eb166ec9-kube-api-access-7jjgr\") pod \"openstackclient\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.245550 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b6fa5609-65ed-4b78-a850-4be3eb166ec9-openstack-config\") pod \"openstackclient\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.245809 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b6fa5609-65ed-4b78-a850-4be3eb166ec9-openstack-config-secret\") pod \"openstackclient\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.245876 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jjgr\" (UniqueName: \"kubernetes.io/projected/b6fa5609-65ed-4b78-a850-4be3eb166ec9-kube-api-access-7jjgr\") pod \"openstackclient\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.245943 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6fa5609-65ed-4b78-a850-4be3eb166ec9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.254305 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b6fa5609-65ed-4b78-a850-4be3eb166ec9-openstack-config-secret\") pod \"openstackclient\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.254331 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6fa5609-65ed-4b78-a850-4be3eb166ec9-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.274564 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b6fa5609-65ed-4b78-a850-4be3eb166ec9-openstack-config\") pod \"openstackclient\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.279971 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jjgr\" (UniqueName: \"kubernetes.io/projected/b6fa5609-65ed-4b78-a850-4be3eb166ec9-kube-api-access-7jjgr\") pod \"openstackclient\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.388981 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.389807 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.399602 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.457094 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.480457 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 02 01:57:26 crc kubenswrapper[4884]: E1202 01:57:26.480920 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" containerName="probe" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.480938 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" containerName="probe" Dec 02 01:57:26 crc kubenswrapper[4884]: E1202 01:57:26.480977 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" containerName="cinder-scheduler" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.480984 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" containerName="cinder-scheduler" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.481145 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" containerName="cinder-scheduler" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.481170 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" containerName="probe" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.481858 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.487721 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 02 01:57:26 crc kubenswrapper[4884]: E1202 01:57:26.538533 4884 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 02 01:57:26 crc kubenswrapper[4884]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_b6fa5609-65ed-4b78-a850-4be3eb166ec9_0(0f6562ad9e2965e3bba8e590ba8e2ec1bf6578121ff0ad5e7b456b2bcb368d40): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"0f6562ad9e2965e3bba8e590ba8e2ec1bf6578121ff0ad5e7b456b2bcb368d40" Netns:"/var/run/netns/179e0924-6298-42b7-81b5-a40428c57827" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=0f6562ad9e2965e3bba8e590ba8e2ec1bf6578121ff0ad5e7b456b2bcb368d40;K8S_POD_UID=b6fa5609-65ed-4b78-a850-4be3eb166ec9" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/b6fa5609-65ed-4b78-a850-4be3eb166ec9]: expected pod UID "b6fa5609-65ed-4b78-a850-4be3eb166ec9" but got "b7e641c0-71a1-4786-aeb9-1fbba21c3e49" from Kube API Dec 02 01:57:26 crc kubenswrapper[4884]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 02 01:57:26 crc kubenswrapper[4884]: > Dec 02 01:57:26 crc kubenswrapper[4884]: E1202 01:57:26.538594 4884 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 02 01:57:26 crc kubenswrapper[4884]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_b6fa5609-65ed-4b78-a850-4be3eb166ec9_0(0f6562ad9e2965e3bba8e590ba8e2ec1bf6578121ff0ad5e7b456b2bcb368d40): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"0f6562ad9e2965e3bba8e590ba8e2ec1bf6578121ff0ad5e7b456b2bcb368d40" Netns:"/var/run/netns/179e0924-6298-42b7-81b5-a40428c57827" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=0f6562ad9e2965e3bba8e590ba8e2ec1bf6578121ff0ad5e7b456b2bcb368d40;K8S_POD_UID=b6fa5609-65ed-4b78-a850-4be3eb166ec9" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/b6fa5609-65ed-4b78-a850-4be3eb166ec9]: expected pod UID "b6fa5609-65ed-4b78-a850-4be3eb166ec9" but got "b7e641c0-71a1-4786-aeb9-1fbba21c3e49" from Kube API Dec 02 01:57:26 crc kubenswrapper[4884]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 02 01:57:26 crc kubenswrapper[4884]: > pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.551241 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-combined-ca-bundle\") pod \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.551698 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-scripts\") pod \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.551991 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-config-data\") pod \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.552039 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-config-data-custom\") pod \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.552087 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzt4t\" (UniqueName: \"kubernetes.io/projected/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-kube-api-access-xzt4t\") pod \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.552180 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-etc-machine-id\") pod \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\" (UID: \"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4\") " Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.552423 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b7e641c0-71a1-4786-aeb9-1fbba21c3e49-openstack-config\") pod \"openstackclient\" (UID: \"b7e641c0-71a1-4786-aeb9-1fbba21c3e49\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.552503 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b7e641c0-71a1-4786-aeb9-1fbba21c3e49-openstack-config-secret\") pod \"openstackclient\" (UID: \"b7e641c0-71a1-4786-aeb9-1fbba21c3e49\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.552558 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fphz5\" (UniqueName: \"kubernetes.io/projected/b7e641c0-71a1-4786-aeb9-1fbba21c3e49-kube-api-access-fphz5\") pod \"openstackclient\" (UID: \"b7e641c0-71a1-4786-aeb9-1fbba21c3e49\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.552586 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e641c0-71a1-4786-aeb9-1fbba21c3e49-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b7e641c0-71a1-4786-aeb9-1fbba21c3e49\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.552724 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" (UID: "df9c4191-0c08-4eaa-9c3f-d5bef64c22e4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.555407 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-scripts" (OuterVolumeSpecName: "scripts") pod "df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" (UID: "df9c4191-0c08-4eaa-9c3f-d5bef64c22e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.555891 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" (UID: "df9c4191-0c08-4eaa-9c3f-d5bef64c22e4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.556594 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-kube-api-access-xzt4t" (OuterVolumeSpecName: "kube-api-access-xzt4t") pod "df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" (UID: "df9c4191-0c08-4eaa-9c3f-d5bef64c22e4"). InnerVolumeSpecName "kube-api-access-xzt4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.609997 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" (UID: "df9c4191-0c08-4eaa-9c3f-d5bef64c22e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.654311 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b7e641c0-71a1-4786-aeb9-1fbba21c3e49-openstack-config-secret\") pod \"openstackclient\" (UID: \"b7e641c0-71a1-4786-aeb9-1fbba21c3e49\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.654397 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fphz5\" (UniqueName: \"kubernetes.io/projected/b7e641c0-71a1-4786-aeb9-1fbba21c3e49-kube-api-access-fphz5\") pod \"openstackclient\" (UID: \"b7e641c0-71a1-4786-aeb9-1fbba21c3e49\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.654421 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e641c0-71a1-4786-aeb9-1fbba21c3e49-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b7e641c0-71a1-4786-aeb9-1fbba21c3e49\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.654503 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b7e641c0-71a1-4786-aeb9-1fbba21c3e49-openstack-config\") pod \"openstackclient\" (UID: \"b7e641c0-71a1-4786-aeb9-1fbba21c3e49\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.654563 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.654573 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.654582 4884 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.654590 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzt4t\" (UniqueName: \"kubernetes.io/projected/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-kube-api-access-xzt4t\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.654601 4884 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.655359 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b7e641c0-71a1-4786-aeb9-1fbba21c3e49-openstack-config\") pod \"openstackclient\" (UID: \"b7e641c0-71a1-4786-aeb9-1fbba21c3e49\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.656434 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-config-data" (OuterVolumeSpecName: "config-data") pod "df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" (UID: "df9c4191-0c08-4eaa-9c3f-d5bef64c22e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.657968 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e641c0-71a1-4786-aeb9-1fbba21c3e49-combined-ca-bundle\") pod \"openstackclient\" (UID: \"b7e641c0-71a1-4786-aeb9-1fbba21c3e49\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.658696 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b7e641c0-71a1-4786-aeb9-1fbba21c3e49-openstack-config-secret\") pod \"openstackclient\" (UID: \"b7e641c0-71a1-4786-aeb9-1fbba21c3e49\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.677477 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fphz5\" (UniqueName: \"kubernetes.io/projected/b7e641c0-71a1-4786-aeb9-1fbba21c3e49-kube-api-access-fphz5\") pod \"openstackclient\" (UID: \"b7e641c0-71a1-4786-aeb9-1fbba21c3e49\") " pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.756579 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.798674 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.872314 4884 generic.go:334] "Generic (PLEG): container finished" podID="df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" containerID="d81a727fe5867ce6a12a75b34235e9466ffee30e83084e6cdc2fb553c25b849e" exitCode=0 Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.872370 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.872387 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4","Type":"ContainerDied","Data":"d81a727fe5867ce6a12a75b34235e9466ffee30e83084e6cdc2fb553c25b849e"} Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.872372 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.872462 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"df9c4191-0c08-4eaa-9c3f-d5bef64c22e4","Type":"ContainerDied","Data":"cd6fc3e4e9929fa45060aa657e328a9079df9aaeb62e1276db94a3a620be1cac"} Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.872501 4884 scope.go:117] "RemoveContainer" containerID="600d4123a3204032be18d4cc42a419e78cb8d87c48eb4a9ce2d3f1b3d1dd6878" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.878316 4884 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="b6fa5609-65ed-4b78-a850-4be3eb166ec9" podUID="b7e641c0-71a1-4786-aeb9-1fbba21c3e49" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.883263 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.925829 4884 scope.go:117] "RemoveContainer" containerID="d81a727fe5867ce6a12a75b34235e9466ffee30e83084e6cdc2fb553c25b849e" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.929621 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.943970 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.961837 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6fa5609-65ed-4b78-a850-4be3eb166ec9-combined-ca-bundle\") pod \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.961979 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b6fa5609-65ed-4b78-a850-4be3eb166ec9-openstack-config\") pod \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.962105 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jjgr\" (UniqueName: \"kubernetes.io/projected/b6fa5609-65ed-4b78-a850-4be3eb166ec9-kube-api-access-7jjgr\") pod \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.962182 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b6fa5609-65ed-4b78-a850-4be3eb166ec9-openstack-config-secret\") pod \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\" (UID: \"b6fa5609-65ed-4b78-a850-4be3eb166ec9\") " Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.963272 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.963832 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6fa5609-65ed-4b78-a850-4be3eb166ec9-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "b6fa5609-65ed-4b78-a850-4be3eb166ec9" (UID: "b6fa5609-65ed-4b78-a850-4be3eb166ec9"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.966073 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.974592 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.975661 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6fa5609-65ed-4b78-a850-4be3eb166ec9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6fa5609-65ed-4b78-a850-4be3eb166ec9" (UID: "b6fa5609-65ed-4b78-a850-4be3eb166ec9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.976503 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6fa5609-65ed-4b78-a850-4be3eb166ec9-kube-api-access-7jjgr" (OuterVolumeSpecName: "kube-api-access-7jjgr") pod "b6fa5609-65ed-4b78-a850-4be3eb166ec9" (UID: "b6fa5609-65ed-4b78-a850-4be3eb166ec9"). InnerVolumeSpecName "kube-api-access-7jjgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.987950 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6fa5609-65ed-4b78-a850-4be3eb166ec9-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "b6fa5609-65ed-4b78-a850-4be3eb166ec9" (UID: "b6fa5609-65ed-4b78-a850-4be3eb166ec9"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.990125 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 02 01:57:26 crc kubenswrapper[4884]: I1202 01:57:26.993913 4884 scope.go:117] "RemoveContainer" containerID="600d4123a3204032be18d4cc42a419e78cb8d87c48eb4a9ce2d3f1b3d1dd6878" Dec 02 01:57:27 crc kubenswrapper[4884]: E1202 01:57:27.004132 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"600d4123a3204032be18d4cc42a419e78cb8d87c48eb4a9ce2d3f1b3d1dd6878\": container with ID starting with 600d4123a3204032be18d4cc42a419e78cb8d87c48eb4a9ce2d3f1b3d1dd6878 not found: ID does not exist" containerID="600d4123a3204032be18d4cc42a419e78cb8d87c48eb4a9ce2d3f1b3d1dd6878" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.004175 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"600d4123a3204032be18d4cc42a419e78cb8d87c48eb4a9ce2d3f1b3d1dd6878"} err="failed to get container status \"600d4123a3204032be18d4cc42a419e78cb8d87c48eb4a9ce2d3f1b3d1dd6878\": rpc error: code = NotFound desc = could not find container \"600d4123a3204032be18d4cc42a419e78cb8d87c48eb4a9ce2d3f1b3d1dd6878\": container with ID starting with 600d4123a3204032be18d4cc42a419e78cb8d87c48eb4a9ce2d3f1b3d1dd6878 not found: ID does not exist" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.004202 4884 scope.go:117] "RemoveContainer" containerID="d81a727fe5867ce6a12a75b34235e9466ffee30e83084e6cdc2fb553c25b849e" Dec 02 01:57:27 crc kubenswrapper[4884]: E1202 01:57:27.004594 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d81a727fe5867ce6a12a75b34235e9466ffee30e83084e6cdc2fb553c25b849e\": container with ID starting with d81a727fe5867ce6a12a75b34235e9466ffee30e83084e6cdc2fb553c25b849e not found: ID does not exist" containerID="d81a727fe5867ce6a12a75b34235e9466ffee30e83084e6cdc2fb553c25b849e" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.004614 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d81a727fe5867ce6a12a75b34235e9466ffee30e83084e6cdc2fb553c25b849e"} err="failed to get container status \"d81a727fe5867ce6a12a75b34235e9466ffee30e83084e6cdc2fb553c25b849e\": rpc error: code = NotFound desc = could not find container \"d81a727fe5867ce6a12a75b34235e9466ffee30e83084e6cdc2fb553c25b849e\": container with ID starting with d81a727fe5867ce6a12a75b34235e9466ffee30e83084e6cdc2fb553c25b849e not found: ID does not exist" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.063796 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfjtq\" (UniqueName: \"kubernetes.io/projected/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-kube-api-access-xfjtq\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.064042 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-config-data\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.064148 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.064252 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-scripts\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.064337 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.064403 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.064512 4884 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b6fa5609-65ed-4b78-a850-4be3eb166ec9-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.064578 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jjgr\" (UniqueName: \"kubernetes.io/projected/b6fa5609-65ed-4b78-a850-4be3eb166ec9-kube-api-access-7jjgr\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.064633 4884 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b6fa5609-65ed-4b78-a850-4be3eb166ec9-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.064685 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6fa5609-65ed-4b78-a850-4be3eb166ec9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.165997 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfjtq\" (UniqueName: \"kubernetes.io/projected/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-kube-api-access-xfjtq\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.166219 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-config-data\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.166333 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.166472 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-scripts\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.166581 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.166658 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.166731 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.170661 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.172354 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.172549 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-config-data\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.182673 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-scripts\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.183322 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfjtq\" (UniqueName: \"kubernetes.io/projected/54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4-kube-api-access-xfjtq\") pod \"cinder-scheduler-0\" (UID: \"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4\") " pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.291416 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.300365 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 02 01:57:27 crc kubenswrapper[4884]: W1202 01:57:27.323560 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7e641c0_71a1_4786_aeb9_1fbba21c3e49.slice/crio-80b17051f16ff15bf8e58f9d1ed4dff8a67f66026c996e09c32e94020d61239e WatchSource:0}: Error finding container 80b17051f16ff15bf8e58f9d1ed4dff8a67f66026c996e09c32e94020d61239e: Status 404 returned error can't find the container with id 80b17051f16ff15bf8e58f9d1ed4dff8a67f66026c996e09c32e94020d61239e Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.628921 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6fa5609-65ed-4b78-a850-4be3eb166ec9" path="/var/lib/kubelet/pods/b6fa5609-65ed-4b78-a850-4be3eb166ec9/volumes" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.630932 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df9c4191-0c08-4eaa-9c3f-d5bef64c22e4" path="/var/lib/kubelet/pods/df9c4191-0c08-4eaa-9c3f-d5bef64c22e4/volumes" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.733125 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.886070 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b7e641c0-71a1-4786-aeb9-1fbba21c3e49","Type":"ContainerStarted","Data":"80b17051f16ff15bf8e58f9d1ed4dff8a67f66026c996e09c32e94020d61239e"} Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.888011 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4","Type":"ContainerStarted","Data":"cbf510ecde13676fcae48948c4fcdb6b871034e1800de1ddffdcfd9db609750d"} Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.889470 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 02 01:57:27 crc kubenswrapper[4884]: I1202 01:57:27.896230 4884 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="b6fa5609-65ed-4b78-a850-4be3eb166ec9" podUID="b7e641c0-71a1-4786-aeb9-1fbba21c3e49" Dec 02 01:57:28 crc kubenswrapper[4884]: I1202 01:57:28.900387 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4","Type":"ContainerStarted","Data":"ff4cef6eba2998a65f683c7c2c61c17cb9e4d9605c8efc786be9444e8da5b0f8"} Dec 02 01:57:28 crc kubenswrapper[4884]: I1202 01:57:28.966867 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 02 01:57:29 crc kubenswrapper[4884]: I1202 01:57:29.140526 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-55fc6c9b58-9v749" podUID="59b51ab4-e0c8-42af-92aa-063e66157aa7" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.159:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.159:8443: connect: connection refused" Dec 02 01:57:29 crc kubenswrapper[4884]: I1202 01:57:29.912677 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4","Type":"ContainerStarted","Data":"7e2a6c742dbc5a784a1aa54d821c5fed14011d061de516f5488431c8cfec2c65"} Dec 02 01:57:29 crc kubenswrapper[4884]: I1202 01:57:29.937448 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.937428437 podStartE2EDuration="3.937428437s" podCreationTimestamp="2025-12-02 01:57:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:57:29.932081065 +0000 UTC m=+1266.607917949" watchObservedRunningTime="2025-12-02 01:57:29.937428437 +0000 UTC m=+1266.613265321" Dec 02 01:57:32 crc kubenswrapper[4884]: I1202 01:57:32.291695 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.277668 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-558bcd5597-pglnv"] Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.279258 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.281166 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.281343 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.283420 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.287008 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-558bcd5597-pglnv"] Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.437497 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzgmj\" (UniqueName: \"kubernetes.io/projected/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-kube-api-access-hzgmj\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.437544 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-config-data\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.437576 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-etc-swift\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.437621 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-combined-ca-bundle\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.438246 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-log-httpd\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.438451 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-run-httpd\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.438510 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-internal-tls-certs\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.438895 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-public-tls-certs\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.520266 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.520635 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="ceilometer-central-agent" containerID="cri-o://3722720b386461ac68ae9a1937a67f53f525d12995f309ceccc98bcd4152dbfa" gracePeriod=30 Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.520690 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="proxy-httpd" containerID="cri-o://744158f31b0164b02b4dd08b62bcd62cda42580407a2a9bed1fd25697ab0e6b1" gracePeriod=30 Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.520726 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="ceilometer-notification-agent" containerID="cri-o://cde32bfe32f8f53ffb99fa8611f3a610e84f6515870159f5cc71820c3071676b" gracePeriod=30 Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.520690 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="sg-core" containerID="cri-o://6bf6a42282d1b7843d02c05bc9da35e24e117c81e0cbc65d9241aed1334871c0" gracePeriod=30 Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.530283 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.180:3000/\": EOF" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.548882 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-internal-tls-certs\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.549095 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-public-tls-certs\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.549200 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzgmj\" (UniqueName: \"kubernetes.io/projected/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-kube-api-access-hzgmj\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.549235 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-config-data\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.549278 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-etc-swift\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.549496 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-combined-ca-bundle\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.549572 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-log-httpd\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.549626 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-run-httpd\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.552578 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-log-httpd\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.557417 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-run-httpd\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.557422 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-etc-swift\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.570669 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-config-data\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.573414 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-combined-ca-bundle\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.573485 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-public-tls-certs\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.574695 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-internal-tls-certs\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.581208 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzgmj\" (UniqueName: \"kubernetes.io/projected/c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c-kube-api-access-hzgmj\") pod \"swift-proxy-558bcd5597-pglnv\" (UID: \"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c\") " pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.653236 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.954534 4884 generic.go:334] "Generic (PLEG): container finished" podID="37b95259-8a7e-43fb-8b52-805528b86103" containerID="744158f31b0164b02b4dd08b62bcd62cda42580407a2a9bed1fd25697ab0e6b1" exitCode=0 Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.954564 4884 generic.go:334] "Generic (PLEG): container finished" podID="37b95259-8a7e-43fb-8b52-805528b86103" containerID="6bf6a42282d1b7843d02c05bc9da35e24e117c81e0cbc65d9241aed1334871c0" exitCode=2 Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.954572 4884 generic.go:334] "Generic (PLEG): container finished" podID="37b95259-8a7e-43fb-8b52-805528b86103" containerID="3722720b386461ac68ae9a1937a67f53f525d12995f309ceccc98bcd4152dbfa" exitCode=0 Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.954591 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37b95259-8a7e-43fb-8b52-805528b86103","Type":"ContainerDied","Data":"744158f31b0164b02b4dd08b62bcd62cda42580407a2a9bed1fd25697ab0e6b1"} Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.954614 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37b95259-8a7e-43fb-8b52-805528b86103","Type":"ContainerDied","Data":"6bf6a42282d1b7843d02c05bc9da35e24e117c81e0cbc65d9241aed1334871c0"} Dec 02 01:57:33 crc kubenswrapper[4884]: I1202 01:57:33.954623 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37b95259-8a7e-43fb-8b52-805528b86103","Type":"ContainerDied","Data":"3722720b386461ac68ae9a1937a67f53f525d12995f309ceccc98bcd4152dbfa"} Dec 02 01:57:37 crc kubenswrapper[4884]: I1202 01:57:37.782983 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 02 01:57:37 crc kubenswrapper[4884]: I1202 01:57:37.996724 4884 generic.go:334] "Generic (PLEG): container finished" podID="37b95259-8a7e-43fb-8b52-805528b86103" containerID="cde32bfe32f8f53ffb99fa8611f3a610e84f6515870159f5cc71820c3071676b" exitCode=0 Dec 02 01:57:37 crc kubenswrapper[4884]: I1202 01:57:37.996822 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37b95259-8a7e-43fb-8b52-805528b86103","Type":"ContainerDied","Data":"cde32bfe32f8f53ffb99fa8611f3a610e84f6515870159f5cc71820c3071676b"} Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.686653 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.771695 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-config-data\") pod \"37b95259-8a7e-43fb-8b52-805528b86103\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.771778 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-sg-core-conf-yaml\") pod \"37b95259-8a7e-43fb-8b52-805528b86103\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.771846 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-combined-ca-bundle\") pod \"37b95259-8a7e-43fb-8b52-805528b86103\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.771915 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37b95259-8a7e-43fb-8b52-805528b86103-log-httpd\") pod \"37b95259-8a7e-43fb-8b52-805528b86103\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.772015 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qm77p\" (UniqueName: \"kubernetes.io/projected/37b95259-8a7e-43fb-8b52-805528b86103-kube-api-access-qm77p\") pod \"37b95259-8a7e-43fb-8b52-805528b86103\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.772081 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-scripts\") pod \"37b95259-8a7e-43fb-8b52-805528b86103\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.772095 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37b95259-8a7e-43fb-8b52-805528b86103-run-httpd\") pod \"37b95259-8a7e-43fb-8b52-805528b86103\" (UID: \"37b95259-8a7e-43fb-8b52-805528b86103\") " Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.773201 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37b95259-8a7e-43fb-8b52-805528b86103-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "37b95259-8a7e-43fb-8b52-805528b86103" (UID: "37b95259-8a7e-43fb-8b52-805528b86103"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.773373 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37b95259-8a7e-43fb-8b52-805528b86103-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "37b95259-8a7e-43fb-8b52-805528b86103" (UID: "37b95259-8a7e-43fb-8b52-805528b86103"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.777554 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37b95259-8a7e-43fb-8b52-805528b86103-kube-api-access-qm77p" (OuterVolumeSpecName: "kube-api-access-qm77p") pod "37b95259-8a7e-43fb-8b52-805528b86103" (UID: "37b95259-8a7e-43fb-8b52-805528b86103"). InnerVolumeSpecName "kube-api-access-qm77p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.779283 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-scripts" (OuterVolumeSpecName: "scripts") pod "37b95259-8a7e-43fb-8b52-805528b86103" (UID: "37b95259-8a7e-43fb-8b52-805528b86103"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.805670 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "37b95259-8a7e-43fb-8b52-805528b86103" (UID: "37b95259-8a7e-43fb-8b52-805528b86103"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.863942 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37b95259-8a7e-43fb-8b52-805528b86103" (UID: "37b95259-8a7e-43fb-8b52-805528b86103"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.876269 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.876303 4884 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37b95259-8a7e-43fb-8b52-805528b86103-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.876313 4884 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.876323 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.876331 4884 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/37b95259-8a7e-43fb-8b52-805528b86103-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.876340 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qm77p\" (UniqueName: \"kubernetes.io/projected/37b95259-8a7e-43fb-8b52-805528b86103-kube-api-access-qm77p\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.889794 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-config-data" (OuterVolumeSpecName: "config-data") pod "37b95259-8a7e-43fb-8b52-805528b86103" (UID: "37b95259-8a7e-43fb-8b52-805528b86103"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.959956 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-558bcd5597-pglnv"] Dec 02 01:57:38 crc kubenswrapper[4884]: I1202 01:57:38.977603 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37b95259-8a7e-43fb-8b52-805528b86103-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.007828 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.007964 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"37b95259-8a7e-43fb-8b52-805528b86103","Type":"ContainerDied","Data":"c93a14b311a2a49d068d3cb80b584bd9eca5abd99260c1846bde4523b5f19397"} Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.008054 4884 scope.go:117] "RemoveContainer" containerID="744158f31b0164b02b4dd08b62bcd62cda42580407a2a9bed1fd25697ab0e6b1" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.010589 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"b7e641c0-71a1-4786-aeb9-1fbba21c3e49","Type":"ContainerStarted","Data":"6e6890e182bf39240a9667a3fcf9fdb4179d6664f996b0aeb1371f0cb36336db"} Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.014777 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-558bcd5597-pglnv" event={"ID":"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c","Type":"ContainerStarted","Data":"45b4ed5befbf48cf4a2dea1cc608f55320c189fc3bd53e9b39e173ae78b819dd"} Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.031975 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.005208223 podStartE2EDuration="13.03195351s" podCreationTimestamp="2025-12-02 01:57:26 +0000 UTC" firstStartedPulling="2025-12-02 01:57:27.327985553 +0000 UTC m=+1264.003822437" lastFinishedPulling="2025-12-02 01:57:38.35473084 +0000 UTC m=+1275.030567724" observedRunningTime="2025-12-02 01:57:39.026188028 +0000 UTC m=+1275.702024922" watchObservedRunningTime="2025-12-02 01:57:39.03195351 +0000 UTC m=+1275.707790414" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.035568 4884 scope.go:117] "RemoveContainer" containerID="6bf6a42282d1b7843d02c05bc9da35e24e117c81e0cbc65d9241aed1334871c0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.072120 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.073114 4884 scope.go:117] "RemoveContainer" containerID="cde32bfe32f8f53ffb99fa8611f3a610e84f6515870159f5cc71820c3071676b" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.088304 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.099398 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:39 crc kubenswrapper[4884]: E1202 01:57:39.099907 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="ceilometer-notification-agent" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.099991 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="ceilometer-notification-agent" Dec 02 01:57:39 crc kubenswrapper[4884]: E1202 01:57:39.100069 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="ceilometer-central-agent" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.100129 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="ceilometer-central-agent" Dec 02 01:57:39 crc kubenswrapper[4884]: E1202 01:57:39.100203 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="sg-core" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.100255 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="sg-core" Dec 02 01:57:39 crc kubenswrapper[4884]: E1202 01:57:39.100312 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="proxy-httpd" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.100364 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="proxy-httpd" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.100609 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="ceilometer-central-agent" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.100672 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="proxy-httpd" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.100730 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="sg-core" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.100812 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="37b95259-8a7e-43fb-8b52-805528b86103" containerName="ceilometer-notification-agent" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.104361 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.108366 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.108598 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.112641 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.126873 4884 scope.go:117] "RemoveContainer" containerID="3722720b386461ac68ae9a1937a67f53f525d12995f309ceccc98bcd4152dbfa" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.135868 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-55fc6c9b58-9v749" podUID="59b51ab4-e0c8-42af-92aa-063e66157aa7" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.159:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.159:8443: connect: connection refused" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.135955 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.182634 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62832812-79b2-4c5c-b32b-65adcffe2111-log-httpd\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.182683 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62832812-79b2-4c5c-b32b-65adcffe2111-run-httpd\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.182807 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-config-data\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.182855 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.182887 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcsc4\" (UniqueName: \"kubernetes.io/projected/62832812-79b2-4c5c-b32b-65adcffe2111-kube-api-access-jcsc4\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.182901 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.182935 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-scripts\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.284703 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcsc4\" (UniqueName: \"kubernetes.io/projected/62832812-79b2-4c5c-b32b-65adcffe2111-kube-api-access-jcsc4\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.284761 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.284795 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-scripts\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.284888 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62832812-79b2-4c5c-b32b-65adcffe2111-log-httpd\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.284906 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62832812-79b2-4c5c-b32b-65adcffe2111-run-httpd\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.284950 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-config-data\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.284982 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.285427 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62832812-79b2-4c5c-b32b-65adcffe2111-run-httpd\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.285677 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62832812-79b2-4c5c-b32b-65adcffe2111-log-httpd\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.288647 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-scripts\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.289617 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.289820 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-config-data\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.293345 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.307947 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcsc4\" (UniqueName: \"kubernetes.io/projected/62832812-79b2-4c5c-b32b-65adcffe2111-kube-api-access-jcsc4\") pod \"ceilometer-0\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.415247 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.415923 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.628658 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37b95259-8a7e-43fb-8b52-805528b86103" path="/var/lib/kubelet/pods/37b95259-8a7e-43fb-8b52-805528b86103/volumes" Dec 02 01:57:39 crc kubenswrapper[4884]: W1202 01:57:39.947812 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62832812_79b2_4c5c_b32b_65adcffe2111.slice/crio-393fb9d125a469a409b639ca6ba9ca00e968c4e4c532f2b5c1e6bc18f672ed28 WatchSource:0}: Error finding container 393fb9d125a469a409b639ca6ba9ca00e968c4e4c532f2b5c1e6bc18f672ed28: Status 404 returned error can't find the container with id 393fb9d125a469a409b639ca6ba9ca00e968c4e4c532f2b5c1e6bc18f672ed28 Dec 02 01:57:39 crc kubenswrapper[4884]: I1202 01:57:39.951509 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:40 crc kubenswrapper[4884]: I1202 01:57:40.049305 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62832812-79b2-4c5c-b32b-65adcffe2111","Type":"ContainerStarted","Data":"393fb9d125a469a409b639ca6ba9ca00e968c4e4c532f2b5c1e6bc18f672ed28"} Dec 02 01:57:40 crc kubenswrapper[4884]: I1202 01:57:40.051302 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-558bcd5597-pglnv" event={"ID":"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c","Type":"ContainerStarted","Data":"dfc5db4bd783eac32289c472948db9f2b7b9f7a99dae935e8bbe8107bc177884"} Dec 02 01:57:40 crc kubenswrapper[4884]: I1202 01:57:40.051347 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-558bcd5597-pglnv" event={"ID":"c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c","Type":"ContainerStarted","Data":"f0d7ecdc625d34c3e2c14440199365bde85c923c666a4f13304a8ae0cad8708b"} Dec 02 01:57:40 crc kubenswrapper[4884]: I1202 01:57:40.051643 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:40 crc kubenswrapper[4884]: I1202 01:57:40.077589 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-558bcd5597-pglnv" podStartSLOduration=7.0775721560000004 podStartE2EDuration="7.077572156s" podCreationTimestamp="2025-12-02 01:57:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:57:40.068106342 +0000 UTC m=+1276.743943226" watchObservedRunningTime="2025-12-02 01:57:40.077572156 +0000 UTC m=+1276.753409040" Dec 02 01:57:41 crc kubenswrapper[4884]: I1202 01:57:41.060108 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:42 crc kubenswrapper[4884]: I1202 01:57:42.070418 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62832812-79b2-4c5c-b32b-65adcffe2111","Type":"ContainerStarted","Data":"53d2289ccbf340e89458c0ee844d3b9d664d61f0fcf820c62b2bfa110439d546"} Dec 02 01:57:43 crc kubenswrapper[4884]: I1202 01:57:43.085015 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62832812-79b2-4c5c-b32b-65adcffe2111","Type":"ContainerStarted","Data":"16aca2b4dd400464d5599d69e0c07138eb845dcbead703527c11b46d08653bcc"} Dec 02 01:57:44 crc kubenswrapper[4884]: I1202 01:57:44.198980 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 02 01:57:44 crc kubenswrapper[4884]: I1202 01:57:44.199419 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/watcher-decision-engine-0" podUID="2417b2c6-1407-4e68-aed3-b4d14802bc9e" containerName="watcher-decision-engine" containerID="cri-o://fb94676af85e01bb962d107a2d9516e111f83b880482244f79a8c359fcf6ee8a" gracePeriod=30 Dec 02 01:57:45 crc kubenswrapper[4884]: I1202 01:57:45.107357 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62832812-79b2-4c5c-b32b-65adcffe2111","Type":"ContainerStarted","Data":"2df672ab4242f99902a8a97d8cecb1cf5283be09977d7bebea393c11d4073f0d"} Dec 02 01:57:45 crc kubenswrapper[4884]: I1202 01:57:45.861431 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-6d8tj"] Dec 02 01:57:45 crc kubenswrapper[4884]: I1202 01:57:45.863581 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6d8tj" Dec 02 01:57:45 crc kubenswrapper[4884]: I1202 01:57:45.886328 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-6d8tj"] Dec 02 01:57:45 crc kubenswrapper[4884]: I1202 01:57:45.916245 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7hdj\" (UniqueName: \"kubernetes.io/projected/ac38729e-6367-4c33-9801-5c79cb0c7fb0-kube-api-access-h7hdj\") pod \"nova-api-db-create-6d8tj\" (UID: \"ac38729e-6367-4c33-9801-5c79cb0c7fb0\") " pod="openstack/nova-api-db-create-6d8tj" Dec 02 01:57:45 crc kubenswrapper[4884]: I1202 01:57:45.916324 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac38729e-6367-4c33-9801-5c79cb0c7fb0-operator-scripts\") pod \"nova-api-db-create-6d8tj\" (UID: \"ac38729e-6367-4c33-9801-5c79cb0c7fb0\") " pod="openstack/nova-api-db-create-6d8tj" Dec 02 01:57:45 crc kubenswrapper[4884]: I1202 01:57:45.941081 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-hdx8d"] Dec 02 01:57:45 crc kubenswrapper[4884]: I1202 01:57:45.943279 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hdx8d" Dec 02 01:57:45 crc kubenswrapper[4884]: I1202 01:57:45.950068 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-hdx8d"] Dec 02 01:57:45 crc kubenswrapper[4884]: I1202 01:57:45.959450 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-ba12-account-create-update-szvnb"] Dec 02 01:57:45 crc kubenswrapper[4884]: I1202 01:57:45.960615 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ba12-account-create-update-szvnb" Dec 02 01:57:45 crc kubenswrapper[4884]: I1202 01:57:45.962560 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 02 01:57:45 crc kubenswrapper[4884]: I1202 01:57:45.971203 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-ba12-account-create-update-szvnb"] Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.021030 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7mr2\" (UniqueName: \"kubernetes.io/projected/2e7e4b35-7937-45cf-9268-4c510bb9bef7-kube-api-access-f7mr2\") pod \"nova-cell0-db-create-hdx8d\" (UID: \"2e7e4b35-7937-45cf-9268-4c510bb9bef7\") " pod="openstack/nova-cell0-db-create-hdx8d" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.021077 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lwzs\" (UniqueName: \"kubernetes.io/projected/97a8c2ed-ae93-4ad6-944c-beca1829d059-kube-api-access-2lwzs\") pod \"nova-api-ba12-account-create-update-szvnb\" (UID: \"97a8c2ed-ae93-4ad6-944c-beca1829d059\") " pod="openstack/nova-api-ba12-account-create-update-szvnb" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.021105 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e7e4b35-7937-45cf-9268-4c510bb9bef7-operator-scripts\") pod \"nova-cell0-db-create-hdx8d\" (UID: \"2e7e4b35-7937-45cf-9268-4c510bb9bef7\") " pod="openstack/nova-cell0-db-create-hdx8d" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.021158 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7hdj\" (UniqueName: \"kubernetes.io/projected/ac38729e-6367-4c33-9801-5c79cb0c7fb0-kube-api-access-h7hdj\") pod \"nova-api-db-create-6d8tj\" (UID: \"ac38729e-6367-4c33-9801-5c79cb0c7fb0\") " pod="openstack/nova-api-db-create-6d8tj" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.021326 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac38729e-6367-4c33-9801-5c79cb0c7fb0-operator-scripts\") pod \"nova-api-db-create-6d8tj\" (UID: \"ac38729e-6367-4c33-9801-5c79cb0c7fb0\") " pod="openstack/nova-api-db-create-6d8tj" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.021388 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97a8c2ed-ae93-4ad6-944c-beca1829d059-operator-scripts\") pod \"nova-api-ba12-account-create-update-szvnb\" (UID: \"97a8c2ed-ae93-4ad6-944c-beca1829d059\") " pod="openstack/nova-api-ba12-account-create-update-szvnb" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.022208 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac38729e-6367-4c33-9801-5c79cb0c7fb0-operator-scripts\") pod \"nova-api-db-create-6d8tj\" (UID: \"ac38729e-6367-4c33-9801-5c79cb0c7fb0\") " pod="openstack/nova-api-db-create-6d8tj" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.043536 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7hdj\" (UniqueName: \"kubernetes.io/projected/ac38729e-6367-4c33-9801-5c79cb0c7fb0-kube-api-access-h7hdj\") pod \"nova-api-db-create-6d8tj\" (UID: \"ac38729e-6367-4c33-9801-5c79cb0c7fb0\") " pod="openstack/nova-api-db-create-6d8tj" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.057012 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-5nk2w"] Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.058592 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5nk2w" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.077831 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5nk2w"] Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.126096 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97a8c2ed-ae93-4ad6-944c-beca1829d059-operator-scripts\") pod \"nova-api-ba12-account-create-update-szvnb\" (UID: \"97a8c2ed-ae93-4ad6-944c-beca1829d059\") " pod="openstack/nova-api-ba12-account-create-update-szvnb" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.126182 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7lxn\" (UniqueName: \"kubernetes.io/projected/3389947b-0d8d-443e-8586-e60740329b9b-kube-api-access-v7lxn\") pod \"nova-cell1-db-create-5nk2w\" (UID: \"3389947b-0d8d-443e-8586-e60740329b9b\") " pod="openstack/nova-cell1-db-create-5nk2w" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.126239 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3389947b-0d8d-443e-8586-e60740329b9b-operator-scripts\") pod \"nova-cell1-db-create-5nk2w\" (UID: \"3389947b-0d8d-443e-8586-e60740329b9b\") " pod="openstack/nova-cell1-db-create-5nk2w" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.126379 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7mr2\" (UniqueName: \"kubernetes.io/projected/2e7e4b35-7937-45cf-9268-4c510bb9bef7-kube-api-access-f7mr2\") pod \"nova-cell0-db-create-hdx8d\" (UID: \"2e7e4b35-7937-45cf-9268-4c510bb9bef7\") " pod="openstack/nova-cell0-db-create-hdx8d" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.126413 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lwzs\" (UniqueName: \"kubernetes.io/projected/97a8c2ed-ae93-4ad6-944c-beca1829d059-kube-api-access-2lwzs\") pod \"nova-api-ba12-account-create-update-szvnb\" (UID: \"97a8c2ed-ae93-4ad6-944c-beca1829d059\") " pod="openstack/nova-api-ba12-account-create-update-szvnb" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.126440 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e7e4b35-7937-45cf-9268-4c510bb9bef7-operator-scripts\") pod \"nova-cell0-db-create-hdx8d\" (UID: \"2e7e4b35-7937-45cf-9268-4c510bb9bef7\") " pod="openstack/nova-cell0-db-create-hdx8d" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.126847 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97a8c2ed-ae93-4ad6-944c-beca1829d059-operator-scripts\") pod \"nova-api-ba12-account-create-update-szvnb\" (UID: \"97a8c2ed-ae93-4ad6-944c-beca1829d059\") " pod="openstack/nova-api-ba12-account-create-update-szvnb" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.127196 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e7e4b35-7937-45cf-9268-4c510bb9bef7-operator-scripts\") pod \"nova-cell0-db-create-hdx8d\" (UID: \"2e7e4b35-7937-45cf-9268-4c510bb9bef7\") " pod="openstack/nova-cell0-db-create-hdx8d" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.151346 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7mr2\" (UniqueName: \"kubernetes.io/projected/2e7e4b35-7937-45cf-9268-4c510bb9bef7-kube-api-access-f7mr2\") pod \"nova-cell0-db-create-hdx8d\" (UID: \"2e7e4b35-7937-45cf-9268-4c510bb9bef7\") " pod="openstack/nova-cell0-db-create-hdx8d" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.158072 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-ee48-account-create-update-bg8ph"] Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.174632 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ee48-account-create-update-bg8ph" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.179477 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6d8tj" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.182478 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lwzs\" (UniqueName: \"kubernetes.io/projected/97a8c2ed-ae93-4ad6-944c-beca1829d059-kube-api-access-2lwzs\") pod \"nova-api-ba12-account-create-update-szvnb\" (UID: \"97a8c2ed-ae93-4ad6-944c-beca1829d059\") " pod="openstack/nova-api-ba12-account-create-update-szvnb" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.185067 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.190193 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ee48-account-create-update-bg8ph"] Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.206804 4884 generic.go:334] "Generic (PLEG): container finished" podID="59b51ab4-e0c8-42af-92aa-063e66157aa7" containerID="d1f74407d395e0019fa6a1d81737da09d7d7e55fb3d882e664487b5cbff7aa32" exitCode=137 Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.206843 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55fc6c9b58-9v749" event={"ID":"59b51ab4-e0c8-42af-92aa-063e66157aa7","Type":"ContainerDied","Data":"d1f74407d395e0019fa6a1d81737da09d7d7e55fb3d882e664487b5cbff7aa32"} Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.234272 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7lxn\" (UniqueName: \"kubernetes.io/projected/3389947b-0d8d-443e-8586-e60740329b9b-kube-api-access-v7lxn\") pod \"nova-cell1-db-create-5nk2w\" (UID: \"3389947b-0d8d-443e-8586-e60740329b9b\") " pod="openstack/nova-cell1-db-create-5nk2w" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.234401 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3389947b-0d8d-443e-8586-e60740329b9b-operator-scripts\") pod \"nova-cell1-db-create-5nk2w\" (UID: \"3389947b-0d8d-443e-8586-e60740329b9b\") " pod="openstack/nova-cell1-db-create-5nk2w" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.234569 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0-operator-scripts\") pod \"nova-cell0-ee48-account-create-update-bg8ph\" (UID: \"b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0\") " pod="openstack/nova-cell0-ee48-account-create-update-bg8ph" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.235304 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3389947b-0d8d-443e-8586-e60740329b9b-operator-scripts\") pod \"nova-cell1-db-create-5nk2w\" (UID: \"3389947b-0d8d-443e-8586-e60740329b9b\") " pod="openstack/nova-cell1-db-create-5nk2w" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.243883 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9m59\" (UniqueName: \"kubernetes.io/projected/b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0-kube-api-access-w9m59\") pod \"nova-cell0-ee48-account-create-update-bg8ph\" (UID: \"b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0\") " pod="openstack/nova-cell0-ee48-account-create-update-bg8ph" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.246654 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7lxn\" (UniqueName: \"kubernetes.io/projected/3389947b-0d8d-443e-8586-e60740329b9b-kube-api-access-v7lxn\") pod \"nova-cell1-db-create-5nk2w\" (UID: \"3389947b-0d8d-443e-8586-e60740329b9b\") " pod="openstack/nova-cell1-db-create-5nk2w" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.256578 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hdx8d" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.273633 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ba12-account-create-update-szvnb" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.284410 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5nk2w" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.346824 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0-operator-scripts\") pod \"nova-cell0-ee48-account-create-update-bg8ph\" (UID: \"b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0\") " pod="openstack/nova-cell0-ee48-account-create-update-bg8ph" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.357381 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9m59\" (UniqueName: \"kubernetes.io/projected/b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0-kube-api-access-w9m59\") pod \"nova-cell0-ee48-account-create-update-bg8ph\" (UID: \"b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0\") " pod="openstack/nova-cell0-ee48-account-create-update-bg8ph" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.350840 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0-operator-scripts\") pod \"nova-cell0-ee48-account-create-update-bg8ph\" (UID: \"b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0\") " pod="openstack/nova-cell0-ee48-account-create-update-bg8ph" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.379732 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9m59\" (UniqueName: \"kubernetes.io/projected/b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0-kube-api-access-w9m59\") pod \"nova-cell0-ee48-account-create-update-bg8ph\" (UID: \"b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0\") " pod="openstack/nova-cell0-ee48-account-create-update-bg8ph" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.379883 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-1ad5-account-create-update-qw2ms"] Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.381058 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1ad5-account-create-update-qw2ms" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.386698 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.410846 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1ad5-account-create-update-qw2ms"] Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.459518 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ceaef10-69d2-488f-becb-9683a34a59df-operator-scripts\") pod \"nova-cell1-1ad5-account-create-update-qw2ms\" (UID: \"3ceaef10-69d2-488f-becb-9683a34a59df\") " pod="openstack/nova-cell1-1ad5-account-create-update-qw2ms" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.459563 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r62rr\" (UniqueName: \"kubernetes.io/projected/3ceaef10-69d2-488f-becb-9683a34a59df-kube-api-access-r62rr\") pod \"nova-cell1-1ad5-account-create-update-qw2ms\" (UID: \"3ceaef10-69d2-488f-becb-9683a34a59df\") " pod="openstack/nova-cell1-1ad5-account-create-update-qw2ms" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.571852 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ceaef10-69d2-488f-becb-9683a34a59df-operator-scripts\") pod \"nova-cell1-1ad5-account-create-update-qw2ms\" (UID: \"3ceaef10-69d2-488f-becb-9683a34a59df\") " pod="openstack/nova-cell1-1ad5-account-create-update-qw2ms" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.571908 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r62rr\" (UniqueName: \"kubernetes.io/projected/3ceaef10-69d2-488f-becb-9683a34a59df-kube-api-access-r62rr\") pod \"nova-cell1-1ad5-account-create-update-qw2ms\" (UID: \"3ceaef10-69d2-488f-becb-9683a34a59df\") " pod="openstack/nova-cell1-1ad5-account-create-update-qw2ms" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.572434 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ceaef10-69d2-488f-becb-9683a34a59df-operator-scripts\") pod \"nova-cell1-1ad5-account-create-update-qw2ms\" (UID: \"3ceaef10-69d2-488f-becb-9683a34a59df\") " pod="openstack/nova-cell1-1ad5-account-create-update-qw2ms" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.613435 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r62rr\" (UniqueName: \"kubernetes.io/projected/3ceaef10-69d2-488f-becb-9683a34a59df-kube-api-access-r62rr\") pod \"nova-cell1-1ad5-account-create-update-qw2ms\" (UID: \"3ceaef10-69d2-488f-becb-9683a34a59df\") " pod="openstack/nova-cell1-1ad5-account-create-update-qw2ms" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.623292 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ee48-account-create-update-bg8ph" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.698023 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.738200 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1ad5-account-create-update-qw2ms" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.777075 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-horizon-secret-key\") pod \"59b51ab4-e0c8-42af-92aa-063e66157aa7\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.777131 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59b51ab4-e0c8-42af-92aa-063e66157aa7-logs\") pod \"59b51ab4-e0c8-42af-92aa-063e66157aa7\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.777183 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsbvn\" (UniqueName: \"kubernetes.io/projected/59b51ab4-e0c8-42af-92aa-063e66157aa7-kube-api-access-zsbvn\") pod \"59b51ab4-e0c8-42af-92aa-063e66157aa7\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.777234 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-combined-ca-bundle\") pod \"59b51ab4-e0c8-42af-92aa-063e66157aa7\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.777267 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59b51ab4-e0c8-42af-92aa-063e66157aa7-scripts\") pod \"59b51ab4-e0c8-42af-92aa-063e66157aa7\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.777453 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-horizon-tls-certs\") pod \"59b51ab4-e0c8-42af-92aa-063e66157aa7\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.777474 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/59b51ab4-e0c8-42af-92aa-063e66157aa7-config-data\") pod \"59b51ab4-e0c8-42af-92aa-063e66157aa7\" (UID: \"59b51ab4-e0c8-42af-92aa-063e66157aa7\") " Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.778298 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59b51ab4-e0c8-42af-92aa-063e66157aa7-logs" (OuterVolumeSpecName: "logs") pod "59b51ab4-e0c8-42af-92aa-063e66157aa7" (UID: "59b51ab4-e0c8-42af-92aa-063e66157aa7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.798839 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "59b51ab4-e0c8-42af-92aa-063e66157aa7" (UID: "59b51ab4-e0c8-42af-92aa-063e66157aa7"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.799237 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59b51ab4-e0c8-42af-92aa-063e66157aa7-kube-api-access-zsbvn" (OuterVolumeSpecName: "kube-api-access-zsbvn") pod "59b51ab4-e0c8-42af-92aa-063e66157aa7" (UID: "59b51ab4-e0c8-42af-92aa-063e66157aa7"). InnerVolumeSpecName "kube-api-access-zsbvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.892318 4884 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.893703 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59b51ab4-e0c8-42af-92aa-063e66157aa7-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.893956 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsbvn\" (UniqueName: \"kubernetes.io/projected/59b51ab4-e0c8-42af-92aa-063e66157aa7-kube-api-access-zsbvn\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.910617 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59b51ab4-e0c8-42af-92aa-063e66157aa7" (UID: "59b51ab4-e0c8-42af-92aa-063e66157aa7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.927709 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59b51ab4-e0c8-42af-92aa-063e66157aa7-config-data" (OuterVolumeSpecName: "config-data") pod "59b51ab4-e0c8-42af-92aa-063e66157aa7" (UID: "59b51ab4-e0c8-42af-92aa-063e66157aa7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.942423 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59b51ab4-e0c8-42af-92aa-063e66157aa7-scripts" (OuterVolumeSpecName: "scripts") pod "59b51ab4-e0c8-42af-92aa-063e66157aa7" (UID: "59b51ab4-e0c8-42af-92aa-063e66157aa7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.943315 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-5nk2w"] Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.952491 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "59b51ab4-e0c8-42af-92aa-063e66157aa7" (UID: "59b51ab4-e0c8-42af-92aa-063e66157aa7"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.986403 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-6d8tj"] Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.995815 4884 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.995854 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/59b51ab4-e0c8-42af-92aa-063e66157aa7-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.995864 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59b51ab4-e0c8-42af-92aa-063e66157aa7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:46 crc kubenswrapper[4884]: I1202 01:57:46.995876 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/59b51ab4-e0c8-42af-92aa-063e66157aa7-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.155345 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-ba12-account-create-update-szvnb"] Dec 02 01:57:47 crc kubenswrapper[4884]: W1202 01:57:47.180155 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e7e4b35_7937_45cf_9268_4c510bb9bef7.slice/crio-c7df9e53accd4e98d8fadd62cfba0287fd9b947f1941736fcb9f47aeb2c1a3ac WatchSource:0}: Error finding container c7df9e53accd4e98d8fadd62cfba0287fd9b947f1941736fcb9f47aeb2c1a3ac: Status 404 returned error can't find the container with id c7df9e53accd4e98d8fadd62cfba0287fd9b947f1941736fcb9f47aeb2c1a3ac Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.191395 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-hdx8d"] Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.236400 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hdx8d" event={"ID":"2e7e4b35-7937-45cf-9268-4c510bb9bef7","Type":"ContainerStarted","Data":"c7df9e53accd4e98d8fadd62cfba0287fd9b947f1941736fcb9f47aeb2c1a3ac"} Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.250867 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55fc6c9b58-9v749" event={"ID":"59b51ab4-e0c8-42af-92aa-063e66157aa7","Type":"ContainerDied","Data":"d4f7cd50430eda19403e476a0d7de9664b4688b9b16b133b01d024e84514f2b5"} Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.251021 4884 scope.go:117] "RemoveContainer" containerID="b52f5ae968a87880245e34143fbe5457d552755e5257d2884c65812f7904191b" Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.251092 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55fc6c9b58-9v749" Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.258332 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-ba12-account-create-update-szvnb" event={"ID":"97a8c2ed-ae93-4ad6-944c-beca1829d059","Type":"ContainerStarted","Data":"8831015a5fd8126cab8bba1d034def529445fa5717be20b0846ee634090962c0"} Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.262792 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62832812-79b2-4c5c-b32b-65adcffe2111","Type":"ContainerStarted","Data":"d589da807b5946726d91b376686620ac5660874ecca8b21cd2fc35c783cc2671"} Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.262942 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="ceilometer-central-agent" containerID="cri-o://53d2289ccbf340e89458c0ee844d3b9d664d61f0fcf820c62b2bfa110439d546" gracePeriod=30 Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.263071 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="proxy-httpd" containerID="cri-o://d589da807b5946726d91b376686620ac5660874ecca8b21cd2fc35c783cc2671" gracePeriod=30 Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.263115 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="sg-core" containerID="cri-o://2df672ab4242f99902a8a97d8cecb1cf5283be09977d7bebea393c11d4073f0d" gracePeriod=30 Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.263153 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="ceilometer-notification-agent" containerID="cri-o://16aca2b4dd400464d5599d69e0c07138eb845dcbead703527c11b46d08653bcc" gracePeriod=30 Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.263218 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.275324 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5nk2w" event={"ID":"3389947b-0d8d-443e-8586-e60740329b9b","Type":"ContainerStarted","Data":"fddfe14029e9bbd768f05df644d7a8734c8cc0f2b9bfb23d424df6f94ce18335"} Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.288883 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6d8tj" event={"ID":"ac38729e-6367-4c33-9801-5c79cb0c7fb0","Type":"ContainerStarted","Data":"c88de53fc8ef6b6aef6e0ffb38ae33c1064f8885777f9121ba2c2c20135e7f23"} Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.292903 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.164315526 podStartE2EDuration="8.292883744s" podCreationTimestamp="2025-12-02 01:57:39 +0000 UTC" firstStartedPulling="2025-12-02 01:57:39.950300273 +0000 UTC m=+1276.626137157" lastFinishedPulling="2025-12-02 01:57:46.078868491 +0000 UTC m=+1282.754705375" observedRunningTime="2025-12-02 01:57:47.282497297 +0000 UTC m=+1283.958334181" watchObservedRunningTime="2025-12-02 01:57:47.292883744 +0000 UTC m=+1283.968720628" Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.372135 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-55fc6c9b58-9v749"] Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.385374 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-55fc6c9b58-9v749"] Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.395277 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-ee48-account-create-update-bg8ph"] Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.404993 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1ad5-account-create-update-qw2ms"] Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.477682 4884 scope.go:117] "RemoveContainer" containerID="d1f74407d395e0019fa6a1d81737da09d7d7e55fb3d882e664487b5cbff7aa32" Dec 02 01:57:47 crc kubenswrapper[4884]: W1202 01:57:47.498861 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ceaef10_69d2_488f_becb_9683a34a59df.slice/crio-515d1ae5891d82935f09323ae662ef4f82c140585eb623872ae3355c4921ae94 WatchSource:0}: Error finding container 515d1ae5891d82935f09323ae662ef4f82c140585eb623872ae3355c4921ae94: Status 404 returned error can't find the container with id 515d1ae5891d82935f09323ae662ef4f82c140585eb623872ae3355c4921ae94 Dec 02 01:57:47 crc kubenswrapper[4884]: I1202 01:57:47.632811 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59b51ab4-e0c8-42af-92aa-063e66157aa7" path="/var/lib/kubelet/pods/59b51ab4-e0c8-42af-92aa-063e66157aa7/volumes" Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.297684 4884 generic.go:334] "Generic (PLEG): container finished" podID="62832812-79b2-4c5c-b32b-65adcffe2111" containerID="d589da807b5946726d91b376686620ac5660874ecca8b21cd2fc35c783cc2671" exitCode=0 Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.298087 4884 generic.go:334] "Generic (PLEG): container finished" podID="62832812-79b2-4c5c-b32b-65adcffe2111" containerID="2df672ab4242f99902a8a97d8cecb1cf5283be09977d7bebea393c11d4073f0d" exitCode=2 Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.298096 4884 generic.go:334] "Generic (PLEG): container finished" podID="62832812-79b2-4c5c-b32b-65adcffe2111" containerID="16aca2b4dd400464d5599d69e0c07138eb845dcbead703527c11b46d08653bcc" exitCode=0 Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.297852 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62832812-79b2-4c5c-b32b-65adcffe2111","Type":"ContainerDied","Data":"d589da807b5946726d91b376686620ac5660874ecca8b21cd2fc35c783cc2671"} Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.298147 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62832812-79b2-4c5c-b32b-65adcffe2111","Type":"ContainerDied","Data":"2df672ab4242f99902a8a97d8cecb1cf5283be09977d7bebea393c11d4073f0d"} Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.298157 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62832812-79b2-4c5c-b32b-65adcffe2111","Type":"ContainerDied","Data":"16aca2b4dd400464d5599d69e0c07138eb845dcbead703527c11b46d08653bcc"} Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.301206 4884 generic.go:334] "Generic (PLEG): container finished" podID="b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0" containerID="8c19950dd25d0e1744f58e97543083aae1f42196c49fb027b5a2876fd7de1a8c" exitCode=0 Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.301249 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ee48-account-create-update-bg8ph" event={"ID":"b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0","Type":"ContainerDied","Data":"8c19950dd25d0e1744f58e97543083aae1f42196c49fb027b5a2876fd7de1a8c"} Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.301267 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ee48-account-create-update-bg8ph" event={"ID":"b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0","Type":"ContainerStarted","Data":"0078cde4b8b3b13a9f6bb3920e7f75149bdd648c67a15c651650f2da25954025"} Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.304877 4884 generic.go:334] "Generic (PLEG): container finished" podID="3389947b-0d8d-443e-8586-e60740329b9b" containerID="621548bbd1d025d4426636ee1e58194771645d67a13795b3461964c8de8c90bb" exitCode=0 Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.304940 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5nk2w" event={"ID":"3389947b-0d8d-443e-8586-e60740329b9b","Type":"ContainerDied","Data":"621548bbd1d025d4426636ee1e58194771645d67a13795b3461964c8de8c90bb"} Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.312445 4884 generic.go:334] "Generic (PLEG): container finished" podID="2417b2c6-1407-4e68-aed3-b4d14802bc9e" containerID="fb94676af85e01bb962d107a2d9516e111f83b880482244f79a8c359fcf6ee8a" exitCode=0 Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.312499 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"2417b2c6-1407-4e68-aed3-b4d14802bc9e","Type":"ContainerDied","Data":"fb94676af85e01bb962d107a2d9516e111f83b880482244f79a8c359fcf6ee8a"} Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.312553 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"2417b2c6-1407-4e68-aed3-b4d14802bc9e","Type":"ContainerDied","Data":"0abe34c3f054d82b6f308f9fd37d9d645be0e0251e0dcf4d517b750683040533"} Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.312566 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0abe34c3f054d82b6f308f9fd37d9d645be0e0251e0dcf4d517b750683040533" Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.314287 4884 generic.go:334] "Generic (PLEG): container finished" podID="ac38729e-6367-4c33-9801-5c79cb0c7fb0" containerID="c9222223659b21deb6248a7babbdceefd8c572f73a4ebf8c62a572b4501bc8f5" exitCode=0 Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.314347 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6d8tj" event={"ID":"ac38729e-6367-4c33-9801-5c79cb0c7fb0","Type":"ContainerDied","Data":"c9222223659b21deb6248a7babbdceefd8c572f73a4ebf8c62a572b4501bc8f5"} Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.327468 4884 generic.go:334] "Generic (PLEG): container finished" podID="3ceaef10-69d2-488f-becb-9683a34a59df" containerID="47231ef26f8635d44ac28086955aa3ffd24d73230815cfb6d1cda21d5690f928" exitCode=0 Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.327556 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1ad5-account-create-update-qw2ms" event={"ID":"3ceaef10-69d2-488f-becb-9683a34a59df","Type":"ContainerDied","Data":"47231ef26f8635d44ac28086955aa3ffd24d73230815cfb6d1cda21d5690f928"} Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.327588 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1ad5-account-create-update-qw2ms" event={"ID":"3ceaef10-69d2-488f-becb-9683a34a59df","Type":"ContainerStarted","Data":"515d1ae5891d82935f09323ae662ef4f82c140585eb623872ae3355c4921ae94"} Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.329485 4884 generic.go:334] "Generic (PLEG): container finished" podID="97a8c2ed-ae93-4ad6-944c-beca1829d059" containerID="225ae74ed43023b6769dfe422299340d8638206e1a4ddcd7cae47f5ba49ada29" exitCode=0 Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.329637 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-ba12-account-create-update-szvnb" event={"ID":"97a8c2ed-ae93-4ad6-944c-beca1829d059","Type":"ContainerDied","Data":"225ae74ed43023b6769dfe422299340d8638206e1a4ddcd7cae47f5ba49ada29"} Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.331277 4884 generic.go:334] "Generic (PLEG): container finished" podID="2e7e4b35-7937-45cf-9268-4c510bb9bef7" containerID="4dce329c1ffe419e86daff3110d7126e144734dac89421e0b57e19f1763ddaa9" exitCode=0 Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.331403 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hdx8d" event={"ID":"2e7e4b35-7937-45cf-9268-4c510bb9bef7","Type":"ContainerDied","Data":"4dce329c1ffe419e86daff3110d7126e144734dac89421e0b57e19f1763ddaa9"} Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.372486 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.428793 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s47z\" (UniqueName: \"kubernetes.io/projected/2417b2c6-1407-4e68-aed3-b4d14802bc9e-kube-api-access-2s47z\") pod \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.428924 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-config-data\") pod \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.428980 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-custom-prometheus-ca\") pod \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.429014 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2417b2c6-1407-4e68-aed3-b4d14802bc9e-logs\") pod \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.429207 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-combined-ca-bundle\") pod \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\" (UID: \"2417b2c6-1407-4e68-aed3-b4d14802bc9e\") " Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.437708 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2417b2c6-1407-4e68-aed3-b4d14802bc9e-logs" (OuterVolumeSpecName: "logs") pod "2417b2c6-1407-4e68-aed3-b4d14802bc9e" (UID: "2417b2c6-1407-4e68-aed3-b4d14802bc9e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.477236 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2417b2c6-1407-4e68-aed3-b4d14802bc9e-kube-api-access-2s47z" (OuterVolumeSpecName: "kube-api-access-2s47z") pod "2417b2c6-1407-4e68-aed3-b4d14802bc9e" (UID: "2417b2c6-1407-4e68-aed3-b4d14802bc9e"). InnerVolumeSpecName "kube-api-access-2s47z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.489005 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-custom-prometheus-ca" (OuterVolumeSpecName: "custom-prometheus-ca") pod "2417b2c6-1407-4e68-aed3-b4d14802bc9e" (UID: "2417b2c6-1407-4e68-aed3-b4d14802bc9e"). InnerVolumeSpecName "custom-prometheus-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.526277 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-config-data" (OuterVolumeSpecName: "config-data") pod "2417b2c6-1407-4e68-aed3-b4d14802bc9e" (UID: "2417b2c6-1407-4e68-aed3-b4d14802bc9e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.529033 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2417b2c6-1407-4e68-aed3-b4d14802bc9e" (UID: "2417b2c6-1407-4e68-aed3-b4d14802bc9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.531097 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.531134 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s47z\" (UniqueName: \"kubernetes.io/projected/2417b2c6-1407-4e68-aed3-b4d14802bc9e-kube-api-access-2s47z\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.531149 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.531157 4884 reconciler_common.go:293] "Volume detached for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/2417b2c6-1407-4e68-aed3-b4d14802bc9e-custom-prometheus-ca\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.531166 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2417b2c6-1407-4e68-aed3-b4d14802bc9e-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.662089 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:48 crc kubenswrapper[4884]: I1202 01:57:48.670247 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-558bcd5597-pglnv" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.020175 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.020774 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="fb5702eb-cee9-4049-b3b4-601f8f5e1854" containerName="glance-log" containerID="cri-o://6448129f68147a08fa73ca4e4a2c9db98fbe699e03fcc1e147e3aa27b7eff2ec" gracePeriod=30 Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.020846 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="fb5702eb-cee9-4049-b3b4-601f8f5e1854" containerName="glance-httpd" containerID="cri-o://d437543659450b227db4b3f26ac8d8eeea74224d10402b9ccfcd8185c9ae56be" gracePeriod=30 Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.342997 4884 generic.go:334] "Generic (PLEG): container finished" podID="fb5702eb-cee9-4049-b3b4-601f8f5e1854" containerID="6448129f68147a08fa73ca4e4a2c9db98fbe699e03fcc1e147e3aa27b7eff2ec" exitCode=143 Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.343087 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fb5702eb-cee9-4049-b3b4-601f8f5e1854","Type":"ContainerDied","Data":"6448129f68147a08fa73ca4e4a2c9db98fbe699e03fcc1e147e3aa27b7eff2ec"} Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.343422 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.390968 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.398859 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.411305 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 02 01:57:49 crc kubenswrapper[4884]: E1202 01:57:49.411686 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2417b2c6-1407-4e68-aed3-b4d14802bc9e" containerName="watcher-decision-engine" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.411700 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2417b2c6-1407-4e68-aed3-b4d14802bc9e" containerName="watcher-decision-engine" Dec 02 01:57:49 crc kubenswrapper[4884]: E1202 01:57:49.411715 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b51ab4-e0c8-42af-92aa-063e66157aa7" containerName="horizon" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.411721 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b51ab4-e0c8-42af-92aa-063e66157aa7" containerName="horizon" Dec 02 01:57:49 crc kubenswrapper[4884]: E1202 01:57:49.411759 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59b51ab4-e0c8-42af-92aa-063e66157aa7" containerName="horizon-log" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.411767 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="59b51ab4-e0c8-42af-92aa-063e66157aa7" containerName="horizon-log" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.411951 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="2417b2c6-1407-4e68-aed3-b4d14802bc9e" containerName="watcher-decision-engine" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.411981 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="59b51ab4-e0c8-42af-92aa-063e66157aa7" containerName="horizon-log" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.411990 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="59b51ab4-e0c8-42af-92aa-063e66157aa7" containerName="horizon" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.412541 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.416075 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"watcher-decision-engine-config-data" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.438630 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.549906 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/40a8993e-7365-473f-9dbd-1f34c001714b-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"40a8993e-7365-473f-9dbd-1f34c001714b\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.549966 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40a8993e-7365-473f-9dbd-1f34c001714b-logs\") pod \"watcher-decision-engine-0\" (UID: \"40a8993e-7365-473f-9dbd-1f34c001714b\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.550018 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb5kd\" (UniqueName: \"kubernetes.io/projected/40a8993e-7365-473f-9dbd-1f34c001714b-kube-api-access-sb5kd\") pod \"watcher-decision-engine-0\" (UID: \"40a8993e-7365-473f-9dbd-1f34c001714b\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.550049 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40a8993e-7365-473f-9dbd-1f34c001714b-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"40a8993e-7365-473f-9dbd-1f34c001714b\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.550098 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40a8993e-7365-473f-9dbd-1f34c001714b-config-data\") pod \"watcher-decision-engine-0\" (UID: \"40a8993e-7365-473f-9dbd-1f34c001714b\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.651887 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/40a8993e-7365-473f-9dbd-1f34c001714b-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"40a8993e-7365-473f-9dbd-1f34c001714b\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.651924 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40a8993e-7365-473f-9dbd-1f34c001714b-logs\") pod \"watcher-decision-engine-0\" (UID: \"40a8993e-7365-473f-9dbd-1f34c001714b\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.651959 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb5kd\" (UniqueName: \"kubernetes.io/projected/40a8993e-7365-473f-9dbd-1f34c001714b-kube-api-access-sb5kd\") pod \"watcher-decision-engine-0\" (UID: \"40a8993e-7365-473f-9dbd-1f34c001714b\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.651984 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40a8993e-7365-473f-9dbd-1f34c001714b-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"40a8993e-7365-473f-9dbd-1f34c001714b\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.652016 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40a8993e-7365-473f-9dbd-1f34c001714b-config-data\") pod \"watcher-decision-engine-0\" (UID: \"40a8993e-7365-473f-9dbd-1f34c001714b\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.652735 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/40a8993e-7365-473f-9dbd-1f34c001714b-logs\") pod \"watcher-decision-engine-0\" (UID: \"40a8993e-7365-473f-9dbd-1f34c001714b\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.662332 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40a8993e-7365-473f-9dbd-1f34c001714b-config-data\") pod \"watcher-decision-engine-0\" (UID: \"40a8993e-7365-473f-9dbd-1f34c001714b\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.664867 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"custom-prometheus-ca\" (UniqueName: \"kubernetes.io/secret/40a8993e-7365-473f-9dbd-1f34c001714b-custom-prometheus-ca\") pod \"watcher-decision-engine-0\" (UID: \"40a8993e-7365-473f-9dbd-1f34c001714b\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.665520 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40a8993e-7365-473f-9dbd-1f34c001714b-combined-ca-bundle\") pod \"watcher-decision-engine-0\" (UID: \"40a8993e-7365-473f-9dbd-1f34c001714b\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.666524 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2417b2c6-1407-4e68-aed3-b4d14802bc9e" path="/var/lib/kubelet/pods/2417b2c6-1407-4e68-aed3-b4d14802bc9e/volumes" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.685206 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb5kd\" (UniqueName: \"kubernetes.io/projected/40a8993e-7365-473f-9dbd-1f34c001714b-kube-api-access-sb5kd\") pod \"watcher-decision-engine-0\" (UID: \"40a8993e-7365-473f-9dbd-1f34c001714b\") " pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.749226 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/watcher-decision-engine-0" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.916428 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ba12-account-create-update-szvnb" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.960883 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lwzs\" (UniqueName: \"kubernetes.io/projected/97a8c2ed-ae93-4ad6-944c-beca1829d059-kube-api-access-2lwzs\") pod \"97a8c2ed-ae93-4ad6-944c-beca1829d059\" (UID: \"97a8c2ed-ae93-4ad6-944c-beca1829d059\") " Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.961023 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97a8c2ed-ae93-4ad6-944c-beca1829d059-operator-scripts\") pod \"97a8c2ed-ae93-4ad6-944c-beca1829d059\" (UID: \"97a8c2ed-ae93-4ad6-944c-beca1829d059\") " Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.962325 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97a8c2ed-ae93-4ad6-944c-beca1829d059-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "97a8c2ed-ae93-4ad6-944c-beca1829d059" (UID: "97a8c2ed-ae93-4ad6-944c-beca1829d059"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:49 crc kubenswrapper[4884]: I1202 01:57:49.966976 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97a8c2ed-ae93-4ad6-944c-beca1829d059-kube-api-access-2lwzs" (OuterVolumeSpecName: "kube-api-access-2lwzs") pod "97a8c2ed-ae93-4ad6-944c-beca1829d059" (UID: "97a8c2ed-ae93-4ad6-944c-beca1829d059"). InnerVolumeSpecName "kube-api-access-2lwzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.063958 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lwzs\" (UniqueName: \"kubernetes.io/projected/97a8c2ed-ae93-4ad6-944c-beca1829d059-kube-api-access-2lwzs\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.064209 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97a8c2ed-ae93-4ad6-944c-beca1829d059-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.139637 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1ad5-account-create-update-qw2ms" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.267818 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ceaef10-69d2-488f-becb-9683a34a59df-operator-scripts\") pod \"3ceaef10-69d2-488f-becb-9683a34a59df\" (UID: \"3ceaef10-69d2-488f-becb-9683a34a59df\") " Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.267926 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r62rr\" (UniqueName: \"kubernetes.io/projected/3ceaef10-69d2-488f-becb-9683a34a59df-kube-api-access-r62rr\") pod \"3ceaef10-69d2-488f-becb-9683a34a59df\" (UID: \"3ceaef10-69d2-488f-becb-9683a34a59df\") " Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.268266 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ceaef10-69d2-488f-becb-9683a34a59df-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3ceaef10-69d2-488f-becb-9683a34a59df" (UID: "3ceaef10-69d2-488f-becb-9683a34a59df"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.268948 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ceaef10-69d2-488f-becb-9683a34a59df-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.272058 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ceaef10-69d2-488f-becb-9683a34a59df-kube-api-access-r62rr" (OuterVolumeSpecName: "kube-api-access-r62rr") pod "3ceaef10-69d2-488f-becb-9683a34a59df" (UID: "3ceaef10-69d2-488f-becb-9683a34a59df"). InnerVolumeSpecName "kube-api-access-r62rr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.275973 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6d8tj" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.330213 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5nk2w" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.336537 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hdx8d" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.347226 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ee48-account-create-update-bg8ph" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.353427 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1ad5-account-create-update-qw2ms" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.353424 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1ad5-account-create-update-qw2ms" event={"ID":"3ceaef10-69d2-488f-becb-9683a34a59df","Type":"ContainerDied","Data":"515d1ae5891d82935f09323ae662ef4f82c140585eb623872ae3355c4921ae94"} Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.353668 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="515d1ae5891d82935f09323ae662ef4f82c140585eb623872ae3355c4921ae94" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.355550 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-ba12-account-create-update-szvnb" event={"ID":"97a8c2ed-ae93-4ad6-944c-beca1829d059","Type":"ContainerDied","Data":"8831015a5fd8126cab8bba1d034def529445fa5717be20b0846ee634090962c0"} Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.355595 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8831015a5fd8126cab8bba1d034def529445fa5717be20b0846ee634090962c0" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.355819 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-ba12-account-create-update-szvnb" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.370243 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac38729e-6367-4c33-9801-5c79cb0c7fb0-operator-scripts\") pod \"ac38729e-6367-4c33-9801-5c79cb0c7fb0\" (UID: \"ac38729e-6367-4c33-9801-5c79cb0c7fb0\") " Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.370303 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7hdj\" (UniqueName: \"kubernetes.io/projected/ac38729e-6367-4c33-9801-5c79cb0c7fb0-kube-api-access-h7hdj\") pod \"ac38729e-6367-4c33-9801-5c79cb0c7fb0\" (UID: \"ac38729e-6367-4c33-9801-5c79cb0c7fb0\") " Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.370710 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r62rr\" (UniqueName: \"kubernetes.io/projected/3ceaef10-69d2-488f-becb-9683a34a59df-kube-api-access-r62rr\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.372922 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac38729e-6367-4c33-9801-5c79cb0c7fb0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ac38729e-6367-4c33-9801-5c79cb0c7fb0" (UID: "ac38729e-6367-4c33-9801-5c79cb0c7fb0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.373133 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-ee48-account-create-update-bg8ph" event={"ID":"b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0","Type":"ContainerDied","Data":"0078cde4b8b3b13a9f6bb3920e7f75149bdd648c67a15c651650f2da25954025"} Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.373164 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0078cde4b8b3b13a9f6bb3920e7f75149bdd648c67a15c651650f2da25954025" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.373232 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-ee48-account-create-update-bg8ph" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.377517 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-5nk2w" event={"ID":"3389947b-0d8d-443e-8586-e60740329b9b","Type":"ContainerDied","Data":"fddfe14029e9bbd768f05df644d7a8734c8cc0f2b9bfb23d424df6f94ce18335"} Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.377579 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fddfe14029e9bbd768f05df644d7a8734c8cc0f2b9bfb23d424df6f94ce18335" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.377653 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-5nk2w" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.377897 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac38729e-6367-4c33-9801-5c79cb0c7fb0-kube-api-access-h7hdj" (OuterVolumeSpecName: "kube-api-access-h7hdj") pod "ac38729e-6367-4c33-9801-5c79cb0c7fb0" (UID: "ac38729e-6367-4c33-9801-5c79cb0c7fb0"). InnerVolumeSpecName "kube-api-access-h7hdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.380644 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6d8tj" event={"ID":"ac38729e-6367-4c33-9801-5c79cb0c7fb0","Type":"ContainerDied","Data":"c88de53fc8ef6b6aef6e0ffb38ae33c1064f8885777f9121ba2c2c20135e7f23"} Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.380685 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c88de53fc8ef6b6aef6e0ffb38ae33c1064f8885777f9121ba2c2c20135e7f23" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.380730 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6d8tj" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.397705 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hdx8d" event={"ID":"2e7e4b35-7937-45cf-9268-4c510bb9bef7","Type":"ContainerDied","Data":"c7df9e53accd4e98d8fadd62cfba0287fd9b947f1941736fcb9f47aeb2c1a3ac"} Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.397762 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7df9e53accd4e98d8fadd62cfba0287fd9b947f1941736fcb9f47aeb2c1a3ac" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.397817 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hdx8d" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.469204 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/watcher-decision-engine-0"] Dec 02 01:57:50 crc kubenswrapper[4884]: W1202 01:57:50.471024 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40a8993e_7365_473f_9dbd_1f34c001714b.slice/crio-f29cc8aae44a8d81a56a59674c4db6e008eb3d1120d5cea5b8517d2a228063cf WatchSource:0}: Error finding container f29cc8aae44a8d81a56a59674c4db6e008eb3d1120d5cea5b8517d2a228063cf: Status 404 returned error can't find the container with id f29cc8aae44a8d81a56a59674c4db6e008eb3d1120d5cea5b8517d2a228063cf Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.471513 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9m59\" (UniqueName: \"kubernetes.io/projected/b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0-kube-api-access-w9m59\") pod \"b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0\" (UID: \"b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0\") " Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.471631 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e7e4b35-7937-45cf-9268-4c510bb9bef7-operator-scripts\") pod \"2e7e4b35-7937-45cf-9268-4c510bb9bef7\" (UID: \"2e7e4b35-7937-45cf-9268-4c510bb9bef7\") " Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.471699 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0-operator-scripts\") pod \"b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0\" (UID: \"b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0\") " Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.471759 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7mr2\" (UniqueName: \"kubernetes.io/projected/2e7e4b35-7937-45cf-9268-4c510bb9bef7-kube-api-access-f7mr2\") pod \"2e7e4b35-7937-45cf-9268-4c510bb9bef7\" (UID: \"2e7e4b35-7937-45cf-9268-4c510bb9bef7\") " Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.471798 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3389947b-0d8d-443e-8586-e60740329b9b-operator-scripts\") pod \"3389947b-0d8d-443e-8586-e60740329b9b\" (UID: \"3389947b-0d8d-443e-8586-e60740329b9b\") " Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.471846 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7lxn\" (UniqueName: \"kubernetes.io/projected/3389947b-0d8d-443e-8586-e60740329b9b-kube-api-access-v7lxn\") pod \"3389947b-0d8d-443e-8586-e60740329b9b\" (UID: \"3389947b-0d8d-443e-8586-e60740329b9b\") " Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.472064 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e7e4b35-7937-45cf-9268-4c510bb9bef7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e7e4b35-7937-45cf-9268-4c510bb9bef7" (UID: "2e7e4b35-7937-45cf-9268-4c510bb9bef7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.472277 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac38729e-6367-4c33-9801-5c79cb0c7fb0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.472303 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e7e4b35-7937-45cf-9268-4c510bb9bef7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.472313 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7hdj\" (UniqueName: \"kubernetes.io/projected/ac38729e-6367-4c33-9801-5c79cb0c7fb0-kube-api-access-h7hdj\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.472310 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0" (UID: "b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.472512 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3389947b-0d8d-443e-8586-e60740329b9b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3389947b-0d8d-443e-8586-e60740329b9b" (UID: "3389947b-0d8d-443e-8586-e60740329b9b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.476942 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3389947b-0d8d-443e-8586-e60740329b9b-kube-api-access-v7lxn" (OuterVolumeSpecName: "kube-api-access-v7lxn") pod "3389947b-0d8d-443e-8586-e60740329b9b" (UID: "3389947b-0d8d-443e-8586-e60740329b9b"). InnerVolumeSpecName "kube-api-access-v7lxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.476988 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e7e4b35-7937-45cf-9268-4c510bb9bef7-kube-api-access-f7mr2" (OuterVolumeSpecName: "kube-api-access-f7mr2") pod "2e7e4b35-7937-45cf-9268-4c510bb9bef7" (UID: "2e7e4b35-7937-45cf-9268-4c510bb9bef7"). InnerVolumeSpecName "kube-api-access-f7mr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.477018 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0-kube-api-access-w9m59" (OuterVolumeSpecName: "kube-api-access-w9m59") pod "b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0" (UID: "b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0"). InnerVolumeSpecName "kube-api-access-w9m59". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.578755 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9m59\" (UniqueName: \"kubernetes.io/projected/b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0-kube-api-access-w9m59\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.578969 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.578979 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7mr2\" (UniqueName: \"kubernetes.io/projected/2e7e4b35-7937-45cf-9268-4c510bb9bef7-kube-api-access-f7mr2\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.578989 4884 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3389947b-0d8d-443e-8586-e60740329b9b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.579002 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7lxn\" (UniqueName: \"kubernetes.io/projected/3389947b-0d8d-443e-8586-e60740329b9b-kube-api-access-v7lxn\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.962287 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.962567 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5fdad2f7-f276-458e-96dd-6cc7fda79cac" containerName="glance-log" containerID="cri-o://abe05486c9de991f37c1e7d97c6cddbef0c9116bc13387312390c4b258d677fb" gracePeriod=30 Dec 02 01:57:50 crc kubenswrapper[4884]: I1202 01:57:50.963007 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5fdad2f7-f276-458e-96dd-6cc7fda79cac" containerName="glance-httpd" containerID="cri-o://40818d37cad862df2661fcdde77ac0513e9e9b1c5f7701e916bdaee848403a76" gracePeriod=30 Dec 02 01:57:51 crc kubenswrapper[4884]: I1202 01:57:51.411565 4884 generic.go:334] "Generic (PLEG): container finished" podID="5fdad2f7-f276-458e-96dd-6cc7fda79cac" containerID="abe05486c9de991f37c1e7d97c6cddbef0c9116bc13387312390c4b258d677fb" exitCode=143 Dec 02 01:57:51 crc kubenswrapper[4884]: I1202 01:57:51.411649 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5fdad2f7-f276-458e-96dd-6cc7fda79cac","Type":"ContainerDied","Data":"abe05486c9de991f37c1e7d97c6cddbef0c9116bc13387312390c4b258d677fb"} Dec 02 01:57:51 crc kubenswrapper[4884]: I1202 01:57:51.413733 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"40a8993e-7365-473f-9dbd-1f34c001714b","Type":"ContainerStarted","Data":"737247d030d4392ef53900ff251697a18152dfb4d496bcc228d3a9e03393dcf4"} Dec 02 01:57:51 crc kubenswrapper[4884]: I1202 01:57:51.413769 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/watcher-decision-engine-0" event={"ID":"40a8993e-7365-473f-9dbd-1f34c001714b","Type":"ContainerStarted","Data":"f29cc8aae44a8d81a56a59674c4db6e008eb3d1120d5cea5b8517d2a228063cf"} Dec 02 01:57:51 crc kubenswrapper[4884]: I1202 01:57:51.429801 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/watcher-decision-engine-0" podStartSLOduration=2.429785159 podStartE2EDuration="2.429785159s" podCreationTimestamp="2025-12-02 01:57:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:57:51.428791165 +0000 UTC m=+1288.104628059" watchObservedRunningTime="2025-12-02 01:57:51.429785159 +0000 UTC m=+1288.105622043" Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.452067 4884 generic.go:334] "Generic (PLEG): container finished" podID="fb5702eb-cee9-4049-b3b4-601f8f5e1854" containerID="d437543659450b227db4b3f26ac8d8eeea74224d10402b9ccfcd8185c9ae56be" exitCode=0 Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.452140 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fb5702eb-cee9-4049-b3b4-601f8f5e1854","Type":"ContainerDied","Data":"d437543659450b227db4b3f26ac8d8eeea74224d10402b9ccfcd8185c9ae56be"} Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.848916 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.951655 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-scripts\") pod \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.951704 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-config-data\") pod \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.951729 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-public-tls-certs\") pod \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.951800 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-combined-ca-bundle\") pod \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.951822 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb5702eb-cee9-4049-b3b4-601f8f5e1854-httpd-run\") pod \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.951840 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5702eb-cee9-4049-b3b4-601f8f5e1854-logs\") pod \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.951934 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx5mr\" (UniqueName: \"kubernetes.io/projected/fb5702eb-cee9-4049-b3b4-601f8f5e1854-kube-api-access-zx5mr\") pod \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.952025 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\" (UID: \"fb5702eb-cee9-4049-b3b4-601f8f5e1854\") " Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.953052 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb5702eb-cee9-4049-b3b4-601f8f5e1854-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "fb5702eb-cee9-4049-b3b4-601f8f5e1854" (UID: "fb5702eb-cee9-4049-b3b4-601f8f5e1854"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.957008 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb5702eb-cee9-4049-b3b4-601f8f5e1854-logs" (OuterVolumeSpecName: "logs") pod "fb5702eb-cee9-4049-b3b4-601f8f5e1854" (UID: "fb5702eb-cee9-4049-b3b4-601f8f5e1854"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.960612 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "fb5702eb-cee9-4049-b3b4-601f8f5e1854" (UID: "fb5702eb-cee9-4049-b3b4-601f8f5e1854"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.965919 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-scripts" (OuterVolumeSpecName: "scripts") pod "fb5702eb-cee9-4049-b3b4-601f8f5e1854" (UID: "fb5702eb-cee9-4049-b3b4-601f8f5e1854"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.965932 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb5702eb-cee9-4049-b3b4-601f8f5e1854-kube-api-access-zx5mr" (OuterVolumeSpecName: "kube-api-access-zx5mr") pod "fb5702eb-cee9-4049-b3b4-601f8f5e1854" (UID: "fb5702eb-cee9-4049-b3b4-601f8f5e1854"). InnerVolumeSpecName "kube-api-access-zx5mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:52 crc kubenswrapper[4884]: I1202 01:57:52.981116 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb5702eb-cee9-4049-b3b4-601f8f5e1854" (UID: "fb5702eb-cee9-4049-b3b4-601f8f5e1854"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.004955 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "fb5702eb-cee9-4049-b3b4-601f8f5e1854" (UID: "fb5702eb-cee9-4049-b3b4-601f8f5e1854"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.006049 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-config-data" (OuterVolumeSpecName: "config-data") pod "fb5702eb-cee9-4049-b3b4-601f8f5e1854" (UID: "fb5702eb-cee9-4049-b3b4-601f8f5e1854"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.053842 4884 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fb5702eb-cee9-4049-b3b4-601f8f5e1854-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.053870 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fb5702eb-cee9-4049-b3b4-601f8f5e1854-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.053879 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx5mr\" (UniqueName: \"kubernetes.io/projected/fb5702eb-cee9-4049-b3b4-601f8f5e1854-kube-api-access-zx5mr\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.053912 4884 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.053923 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.053933 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.053942 4884 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.053951 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb5702eb-cee9-4049-b3b4-601f8f5e1854-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.073224 4884 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.155558 4884 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.467192 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"fb5702eb-cee9-4049-b3b4-601f8f5e1854","Type":"ContainerDied","Data":"06a8b59e950e9d35a57f05254d11958b7f2b55b279c3bdc1ddcdbbebfe783b2c"} Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.467257 4884 scope.go:117] "RemoveContainer" containerID="d437543659450b227db4b3f26ac8d8eeea74224d10402b9ccfcd8185c9ae56be" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.467277 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.492462 4884 scope.go:117] "RemoveContainer" containerID="6448129f68147a08fa73ca4e4a2c9db98fbe699e03fcc1e147e3aa27b7eff2ec" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.550916 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.574699 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.582855 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:57:53 crc kubenswrapper[4884]: E1202 01:57:53.583471 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb5702eb-cee9-4049-b3b4-601f8f5e1854" containerName="glance-httpd" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.583501 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb5702eb-cee9-4049-b3b4-601f8f5e1854" containerName="glance-httpd" Dec 02 01:57:53 crc kubenswrapper[4884]: E1202 01:57:53.583521 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ceaef10-69d2-488f-becb-9683a34a59df" containerName="mariadb-account-create-update" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.583530 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ceaef10-69d2-488f-becb-9683a34a59df" containerName="mariadb-account-create-update" Dec 02 01:57:53 crc kubenswrapper[4884]: E1202 01:57:53.583549 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e7e4b35-7937-45cf-9268-4c510bb9bef7" containerName="mariadb-database-create" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.583557 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e7e4b35-7937-45cf-9268-4c510bb9bef7" containerName="mariadb-database-create" Dec 02 01:57:53 crc kubenswrapper[4884]: E1202 01:57:53.583572 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb5702eb-cee9-4049-b3b4-601f8f5e1854" containerName="glance-log" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.583582 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb5702eb-cee9-4049-b3b4-601f8f5e1854" containerName="glance-log" Dec 02 01:57:53 crc kubenswrapper[4884]: E1202 01:57:53.583593 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0" containerName="mariadb-account-create-update" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.583601 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0" containerName="mariadb-account-create-update" Dec 02 01:57:53 crc kubenswrapper[4884]: E1202 01:57:53.583609 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97a8c2ed-ae93-4ad6-944c-beca1829d059" containerName="mariadb-account-create-update" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.583617 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a8c2ed-ae93-4ad6-944c-beca1829d059" containerName="mariadb-account-create-update" Dec 02 01:57:53 crc kubenswrapper[4884]: E1202 01:57:53.583637 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3389947b-0d8d-443e-8586-e60740329b9b" containerName="mariadb-database-create" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.583644 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="3389947b-0d8d-443e-8586-e60740329b9b" containerName="mariadb-database-create" Dec 02 01:57:53 crc kubenswrapper[4884]: E1202 01:57:53.583666 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac38729e-6367-4c33-9801-5c79cb0c7fb0" containerName="mariadb-database-create" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.583674 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac38729e-6367-4c33-9801-5c79cb0c7fb0" containerName="mariadb-database-create" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.583990 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e7e4b35-7937-45cf-9268-4c510bb9bef7" containerName="mariadb-database-create" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.584019 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="97a8c2ed-ae93-4ad6-944c-beca1829d059" containerName="mariadb-account-create-update" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.584033 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ceaef10-69d2-488f-becb-9683a34a59df" containerName="mariadb-account-create-update" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.584045 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb5702eb-cee9-4049-b3b4-601f8f5e1854" containerName="glance-log" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.584081 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0" containerName="mariadb-account-create-update" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.584095 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb5702eb-cee9-4049-b3b4-601f8f5e1854" containerName="glance-httpd" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.584109 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="3389947b-0d8d-443e-8586-e60740329b9b" containerName="mariadb-database-create" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.584127 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac38729e-6367-4c33-9801-5c79cb0c7fb0" containerName="mariadb-database-create" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.585613 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.589385 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.596006 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.601177 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.631472 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb5702eb-cee9-4049-b3b4-601f8f5e1854" path="/var/lib/kubelet/pods/fb5702eb-cee9-4049-b3b4-601f8f5e1854/volumes" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.664199 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18baf695-0c9e-4c27-a116-13257bd7d281-logs\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.664550 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.664681 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxvjx\" (UniqueName: \"kubernetes.io/projected/18baf695-0c9e-4c27-a116-13257bd7d281-kube-api-access-qxvjx\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.664798 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18baf695-0c9e-4c27-a116-13257bd7d281-config-data\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.664877 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18baf695-0c9e-4c27-a116-13257bd7d281-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.665025 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18baf695-0c9e-4c27-a116-13257bd7d281-scripts\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.665127 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/18baf695-0c9e-4c27-a116-13257bd7d281-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.665317 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/18baf695-0c9e-4c27-a116-13257bd7d281-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.767043 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/18baf695-0c9e-4c27-a116-13257bd7d281-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.768241 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18baf695-0c9e-4c27-a116-13257bd7d281-logs\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.768910 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.768737 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18baf695-0c9e-4c27-a116-13257bd7d281-logs\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.767820 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/18baf695-0c9e-4c27-a116-13257bd7d281-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.769282 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.769608 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxvjx\" (UniqueName: \"kubernetes.io/projected/18baf695-0c9e-4c27-a116-13257bd7d281-kube-api-access-qxvjx\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.769786 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18baf695-0c9e-4c27-a116-13257bd7d281-config-data\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.769924 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18baf695-0c9e-4c27-a116-13257bd7d281-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.770054 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18baf695-0c9e-4c27-a116-13257bd7d281-scripts\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.770234 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/18baf695-0c9e-4c27-a116-13257bd7d281-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.801702 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18baf695-0c9e-4c27-a116-13257bd7d281-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.801931 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxvjx\" (UniqueName: \"kubernetes.io/projected/18baf695-0c9e-4c27-a116-13257bd7d281-kube-api-access-qxvjx\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.803074 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/18baf695-0c9e-4c27-a116-13257bd7d281-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.803359 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18baf695-0c9e-4c27-a116-13257bd7d281-config-data\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.811261 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18baf695-0c9e-4c27-a116-13257bd7d281-scripts\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.822908 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"18baf695-0c9e-4c27-a116-13257bd7d281\") " pod="openstack/glance-default-external-api-0" Dec 02 01:57:53 crc kubenswrapper[4884]: I1202 01:57:53.970510 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.084906 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.184914 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-scripts\") pod \"62832812-79b2-4c5c-b32b-65adcffe2111\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.185033 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62832812-79b2-4c5c-b32b-65adcffe2111-log-httpd\") pod \"62832812-79b2-4c5c-b32b-65adcffe2111\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.185065 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcsc4\" (UniqueName: \"kubernetes.io/projected/62832812-79b2-4c5c-b32b-65adcffe2111-kube-api-access-jcsc4\") pod \"62832812-79b2-4c5c-b32b-65adcffe2111\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.185188 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-combined-ca-bundle\") pod \"62832812-79b2-4c5c-b32b-65adcffe2111\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.185211 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62832812-79b2-4c5c-b32b-65adcffe2111-run-httpd\") pod \"62832812-79b2-4c5c-b32b-65adcffe2111\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.185240 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-sg-core-conf-yaml\") pod \"62832812-79b2-4c5c-b32b-65adcffe2111\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.185258 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-config-data\") pod \"62832812-79b2-4c5c-b32b-65adcffe2111\" (UID: \"62832812-79b2-4c5c-b32b-65adcffe2111\") " Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.185629 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62832812-79b2-4c5c-b32b-65adcffe2111-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "62832812-79b2-4c5c-b32b-65adcffe2111" (UID: "62832812-79b2-4c5c-b32b-65adcffe2111"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.186026 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62832812-79b2-4c5c-b32b-65adcffe2111-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "62832812-79b2-4c5c-b32b-65adcffe2111" (UID: "62832812-79b2-4c5c-b32b-65adcffe2111"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.202113 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62832812-79b2-4c5c-b32b-65adcffe2111-kube-api-access-jcsc4" (OuterVolumeSpecName: "kube-api-access-jcsc4") pod "62832812-79b2-4c5c-b32b-65adcffe2111" (UID: "62832812-79b2-4c5c-b32b-65adcffe2111"). InnerVolumeSpecName "kube-api-access-jcsc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.204180 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-scripts" (OuterVolumeSpecName: "scripts") pod "62832812-79b2-4c5c-b32b-65adcffe2111" (UID: "62832812-79b2-4c5c-b32b-65adcffe2111"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.232577 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "62832812-79b2-4c5c-b32b-65adcffe2111" (UID: "62832812-79b2-4c5c-b32b-65adcffe2111"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.260625 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62832812-79b2-4c5c-b32b-65adcffe2111" (UID: "62832812-79b2-4c5c-b32b-65adcffe2111"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.287420 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.287452 4884 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62832812-79b2-4c5c-b32b-65adcffe2111-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.287461 4884 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.287469 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.287477 4884 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/62832812-79b2-4c5c-b32b-65adcffe2111-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.287488 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcsc4\" (UniqueName: \"kubernetes.io/projected/62832812-79b2-4c5c-b32b-65adcffe2111-kube-api-access-jcsc4\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.305951 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-config-data" (OuterVolumeSpecName: "config-data") pod "62832812-79b2-4c5c-b32b-65adcffe2111" (UID: "62832812-79b2-4c5c-b32b-65adcffe2111"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.389259 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62832812-79b2-4c5c-b32b-65adcffe2111-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.482206 4884 generic.go:334] "Generic (PLEG): container finished" podID="62832812-79b2-4c5c-b32b-65adcffe2111" containerID="53d2289ccbf340e89458c0ee844d3b9d664d61f0fcf820c62b2bfa110439d546" exitCode=0 Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.482305 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62832812-79b2-4c5c-b32b-65adcffe2111","Type":"ContainerDied","Data":"53d2289ccbf340e89458c0ee844d3b9d664d61f0fcf820c62b2bfa110439d546"} Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.482342 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"62832812-79b2-4c5c-b32b-65adcffe2111","Type":"ContainerDied","Data":"393fb9d125a469a409b639ca6ba9ca00e968c4e4c532f2b5c1e6bc18f672ed28"} Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.482338 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.482359 4884 scope.go:117] "RemoveContainer" containerID="d589da807b5946726d91b376686620ac5660874ecca8b21cd2fc35c783cc2671" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.485245 4884 generic.go:334] "Generic (PLEG): container finished" podID="5fdad2f7-f276-458e-96dd-6cc7fda79cac" containerID="40818d37cad862df2661fcdde77ac0513e9e9b1c5f7701e916bdaee848403a76" exitCode=0 Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.485289 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5fdad2f7-f276-458e-96dd-6cc7fda79cac","Type":"ContainerDied","Data":"40818d37cad862df2661fcdde77ac0513e9e9b1c5f7701e916bdaee848403a76"} Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.526163 4884 scope.go:117] "RemoveContainer" containerID="2df672ab4242f99902a8a97d8cecb1cf5283be09977d7bebea393c11d4073f0d" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.527203 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.545719 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.575970 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.596397 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:54 crc kubenswrapper[4884]: E1202 01:57:54.596836 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="ceilometer-central-agent" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.596852 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="ceilometer-central-agent" Dec 02 01:57:54 crc kubenswrapper[4884]: E1202 01:57:54.596871 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="sg-core" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.596877 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="sg-core" Dec 02 01:57:54 crc kubenswrapper[4884]: E1202 01:57:54.596894 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="proxy-httpd" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.596902 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="proxy-httpd" Dec 02 01:57:54 crc kubenswrapper[4884]: E1202 01:57:54.596924 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="ceilometer-notification-agent" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.596930 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="ceilometer-notification-agent" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.597105 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="proxy-httpd" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.597122 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="ceilometer-notification-agent" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.597136 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="ceilometer-central-agent" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.597148 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" containerName="sg-core" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.599230 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.602928 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.603366 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.614960 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.673706 4884 scope.go:117] "RemoveContainer" containerID="16aca2b4dd400464d5599d69e0c07138eb845dcbead703527c11b46d08653bcc" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.695610 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-config-data\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.695659 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.695686 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-scripts\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.695712 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f07d30e-bd91-4bff-bcc2-839db028de27-run-httpd\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.695801 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.695866 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f07d30e-bd91-4bff-bcc2-839db028de27-log-httpd\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.695883 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tczdp\" (UniqueName: \"kubernetes.io/projected/7f07d30e-bd91-4bff-bcc2-839db028de27-kube-api-access-tczdp\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.709549 4884 scope.go:117] "RemoveContainer" containerID="53d2289ccbf340e89458c0ee844d3b9d664d61f0fcf820c62b2bfa110439d546" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.802924 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.803477 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f07d30e-bd91-4bff-bcc2-839db028de27-log-httpd\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.803505 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tczdp\" (UniqueName: \"kubernetes.io/projected/7f07d30e-bd91-4bff-bcc2-839db028de27-kube-api-access-tczdp\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.803550 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-config-data\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.803573 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.803595 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-scripts\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.803623 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f07d30e-bd91-4bff-bcc2-839db028de27-run-httpd\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.804710 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f07d30e-bd91-4bff-bcc2-839db028de27-run-httpd\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.806519 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f07d30e-bd91-4bff-bcc2-839db028de27-log-httpd\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.809801 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.810173 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-config-data\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.816828 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-scripts\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.816945 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.825670 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tczdp\" (UniqueName: \"kubernetes.io/projected/7f07d30e-bd91-4bff-bcc2-839db028de27-kube-api-access-tczdp\") pod \"ceilometer-0\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.826828 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.832580 4884 scope.go:117] "RemoveContainer" containerID="d589da807b5946726d91b376686620ac5660874ecca8b21cd2fc35c783cc2671" Dec 02 01:57:54 crc kubenswrapper[4884]: E1202 01:57:54.833333 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d589da807b5946726d91b376686620ac5660874ecca8b21cd2fc35c783cc2671\": container with ID starting with d589da807b5946726d91b376686620ac5660874ecca8b21cd2fc35c783cc2671 not found: ID does not exist" containerID="d589da807b5946726d91b376686620ac5660874ecca8b21cd2fc35c783cc2671" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.833370 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d589da807b5946726d91b376686620ac5660874ecca8b21cd2fc35c783cc2671"} err="failed to get container status \"d589da807b5946726d91b376686620ac5660874ecca8b21cd2fc35c783cc2671\": rpc error: code = NotFound desc = could not find container \"d589da807b5946726d91b376686620ac5660874ecca8b21cd2fc35c783cc2671\": container with ID starting with d589da807b5946726d91b376686620ac5660874ecca8b21cd2fc35c783cc2671 not found: ID does not exist" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.833392 4884 scope.go:117] "RemoveContainer" containerID="2df672ab4242f99902a8a97d8cecb1cf5283be09977d7bebea393c11d4073f0d" Dec 02 01:57:54 crc kubenswrapper[4884]: E1202 01:57:54.833833 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2df672ab4242f99902a8a97d8cecb1cf5283be09977d7bebea393c11d4073f0d\": container with ID starting with 2df672ab4242f99902a8a97d8cecb1cf5283be09977d7bebea393c11d4073f0d not found: ID does not exist" containerID="2df672ab4242f99902a8a97d8cecb1cf5283be09977d7bebea393c11d4073f0d" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.833872 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2df672ab4242f99902a8a97d8cecb1cf5283be09977d7bebea393c11d4073f0d"} err="failed to get container status \"2df672ab4242f99902a8a97d8cecb1cf5283be09977d7bebea393c11d4073f0d\": rpc error: code = NotFound desc = could not find container \"2df672ab4242f99902a8a97d8cecb1cf5283be09977d7bebea393c11d4073f0d\": container with ID starting with 2df672ab4242f99902a8a97d8cecb1cf5283be09977d7bebea393c11d4073f0d not found: ID does not exist" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.833907 4884 scope.go:117] "RemoveContainer" containerID="16aca2b4dd400464d5599d69e0c07138eb845dcbead703527c11b46d08653bcc" Dec 02 01:57:54 crc kubenswrapper[4884]: E1202 01:57:54.836339 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16aca2b4dd400464d5599d69e0c07138eb845dcbead703527c11b46d08653bcc\": container with ID starting with 16aca2b4dd400464d5599d69e0c07138eb845dcbead703527c11b46d08653bcc not found: ID does not exist" containerID="16aca2b4dd400464d5599d69e0c07138eb845dcbead703527c11b46d08653bcc" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.836363 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16aca2b4dd400464d5599d69e0c07138eb845dcbead703527c11b46d08653bcc"} err="failed to get container status \"16aca2b4dd400464d5599d69e0c07138eb845dcbead703527c11b46d08653bcc\": rpc error: code = NotFound desc = could not find container \"16aca2b4dd400464d5599d69e0c07138eb845dcbead703527c11b46d08653bcc\": container with ID starting with 16aca2b4dd400464d5599d69e0c07138eb845dcbead703527c11b46d08653bcc not found: ID does not exist" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.836382 4884 scope.go:117] "RemoveContainer" containerID="53d2289ccbf340e89458c0ee844d3b9d664d61f0fcf820c62b2bfa110439d546" Dec 02 01:57:54 crc kubenswrapper[4884]: E1202 01:57:54.843917 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53d2289ccbf340e89458c0ee844d3b9d664d61f0fcf820c62b2bfa110439d546\": container with ID starting with 53d2289ccbf340e89458c0ee844d3b9d664d61f0fcf820c62b2bfa110439d546 not found: ID does not exist" containerID="53d2289ccbf340e89458c0ee844d3b9d664d61f0fcf820c62b2bfa110439d546" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.843981 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53d2289ccbf340e89458c0ee844d3b9d664d61f0fcf820c62b2bfa110439d546"} err="failed to get container status \"53d2289ccbf340e89458c0ee844d3b9d664d61f0fcf820c62b2bfa110439d546\": rpc error: code = NotFound desc = could not find container \"53d2289ccbf340e89458c0ee844d3b9d664d61f0fcf820c62b2bfa110439d546\": container with ID starting with 53d2289ccbf340e89458c0ee844d3b9d664d61f0fcf820c62b2bfa110439d546 not found: ID does not exist" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.904411 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fdad2f7-f276-458e-96dd-6cc7fda79cac-logs\") pod \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.904492 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-combined-ca-bundle\") pod \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.904515 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5fdad2f7-f276-458e-96dd-6cc7fda79cac-httpd-run\") pod \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.904584 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-config-data\") pod \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.904601 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-scripts\") pod \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.904631 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.904676 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27fl9\" (UniqueName: \"kubernetes.io/projected/5fdad2f7-f276-458e-96dd-6cc7fda79cac-kube-api-access-27fl9\") pod \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.904769 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-internal-tls-certs\") pod \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\" (UID: \"5fdad2f7-f276-458e-96dd-6cc7fda79cac\") " Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.910032 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "5fdad2f7-f276-458e-96dd-6cc7fda79cac" (UID: "5fdad2f7-f276-458e-96dd-6cc7fda79cac"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.910673 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fdad2f7-f276-458e-96dd-6cc7fda79cac-logs" (OuterVolumeSpecName: "logs") pod "5fdad2f7-f276-458e-96dd-6cc7fda79cac" (UID: "5fdad2f7-f276-458e-96dd-6cc7fda79cac"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.912938 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fdad2f7-f276-458e-96dd-6cc7fda79cac-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5fdad2f7-f276-458e-96dd-6cc7fda79cac" (UID: "5fdad2f7-f276-458e-96dd-6cc7fda79cac"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.917355 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fdad2f7-f276-458e-96dd-6cc7fda79cac-kube-api-access-27fl9" (OuterVolumeSpecName: "kube-api-access-27fl9") pod "5fdad2f7-f276-458e-96dd-6cc7fda79cac" (UID: "5fdad2f7-f276-458e-96dd-6cc7fda79cac"). InnerVolumeSpecName "kube-api-access-27fl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.917587 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-scripts" (OuterVolumeSpecName: "scripts") pod "5fdad2f7-f276-458e-96dd-6cc7fda79cac" (UID: "5fdad2f7-f276-458e-96dd-6cc7fda79cac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.930681 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:57:54 crc kubenswrapper[4884]: I1202 01:57:54.970124 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5fdad2f7-f276-458e-96dd-6cc7fda79cac" (UID: "5fdad2f7-f276-458e-96dd-6cc7fda79cac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.004221 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5fdad2f7-f276-458e-96dd-6cc7fda79cac" (UID: "5fdad2f7-f276-458e-96dd-6cc7fda79cac"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.008566 4884 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5fdad2f7-f276-458e-96dd-6cc7fda79cac-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.008601 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.008630 4884 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.008641 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27fl9\" (UniqueName: \"kubernetes.io/projected/5fdad2f7-f276-458e-96dd-6cc7fda79cac-kube-api-access-27fl9\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.008653 4884 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.008661 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fdad2f7-f276-458e-96dd-6cc7fda79cac-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.008668 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.024956 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-config-data" (OuterVolumeSpecName: "config-data") pod "5fdad2f7-f276-458e-96dd-6cc7fda79cac" (UID: "5fdad2f7-f276-458e-96dd-6cc7fda79cac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.041695 4884 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.110165 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fdad2f7-f276-458e-96dd-6cc7fda79cac-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.110619 4884 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.451123 4884 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.453203 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.543412 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f07d30e-bd91-4bff-bcc2-839db028de27","Type":"ContainerStarted","Data":"a159f7916573de4e3d20ba4dd94bd17d3fcbd46c36865c29be69dd1ea38fb465"} Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.563242 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5fdad2f7-f276-458e-96dd-6cc7fda79cac","Type":"ContainerDied","Data":"6ce0a59277c766ef0cb22193f407931d078d6a831472bf42e83ab6d2d475f76e"} Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.563298 4884 scope.go:117] "RemoveContainer" containerID="40818d37cad862df2661fcdde77ac0513e9e9b1c5f7701e916bdaee848403a76" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.563451 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.597865 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"18baf695-0c9e-4c27-a116-13257bd7d281","Type":"ContainerStarted","Data":"9a935d7890b4f63ee640f436149ba4e0266f154786f032da9e8eeb13b5a14a13"} Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.651733 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62832812-79b2-4c5c-b32b-65adcffe2111" path="/var/lib/kubelet/pods/62832812-79b2-4c5c-b32b-65adcffe2111/volumes" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.732027 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.739131 4884 scope.go:117] "RemoveContainer" containerID="abe05486c9de991f37c1e7d97c6cddbef0c9116bc13387312390c4b258d677fb" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.745656 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.752375 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:57:55 crc kubenswrapper[4884]: E1202 01:57:55.752886 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fdad2f7-f276-458e-96dd-6cc7fda79cac" containerName="glance-httpd" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.752906 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fdad2f7-f276-458e-96dd-6cc7fda79cac" containerName="glance-httpd" Dec 02 01:57:55 crc kubenswrapper[4884]: E1202 01:57:55.752927 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fdad2f7-f276-458e-96dd-6cc7fda79cac" containerName="glance-log" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.752936 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fdad2f7-f276-458e-96dd-6cc7fda79cac" containerName="glance-log" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.753250 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fdad2f7-f276-458e-96dd-6cc7fda79cac" containerName="glance-httpd" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.753273 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fdad2f7-f276-458e-96dd-6cc7fda79cac" containerName="glance-log" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.755862 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.758896 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.758949 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.776697 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.933651 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.933977 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.934007 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-logs\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.934062 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.934114 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.934131 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.934155 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvp2p\" (UniqueName: \"kubernetes.io/projected/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-kube-api-access-gvp2p\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:55 crc kubenswrapper[4884]: I1202 01:57:55.934233 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.035635 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.036376 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.036407 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.036433 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvp2p\" (UniqueName: \"kubernetes.io/projected/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-kube-api-access-gvp2p\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.036496 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.036533 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.036571 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.036591 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-logs\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.036824 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.037269 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-logs\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.037307 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.041733 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.042213 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.042265 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.043695 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.051738 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvp2p\" (UniqueName: \"kubernetes.io/projected/d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c-kube-api-access-gvp2p\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.069288 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c\") " pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.093364 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.429798 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-f48ww"] Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.431527 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.434707 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-52l2h" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.438230 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.441992 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.451955 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-f48ww"] Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.547627 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-f48ww\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.547704 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-config-data\") pod \"nova-cell0-conductor-db-sync-f48ww\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.547817 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-scripts\") pod \"nova-cell0-conductor-db-sync-f48ww\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.547903 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfxz4\" (UniqueName: \"kubernetes.io/projected/26734dfd-6dba-4c5f-99fc-bc27be1b8792-kube-api-access-lfxz4\") pod \"nova-cell0-conductor-db-sync-f48ww\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.626572 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"18baf695-0c9e-4c27-a116-13257bd7d281","Type":"ContainerStarted","Data":"8e0b715d8a01b4d293f09211e75f2e379e0cce8d128493628e43a95fda205f4b"} Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.633754 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"18baf695-0c9e-4c27-a116-13257bd7d281","Type":"ContainerStarted","Data":"509fafe84bf8dcecdce582cb85a211f1aec884e0306328d8fe4c9c5bb7b69c59"} Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.636632 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f07d30e-bd91-4bff-bcc2-839db028de27","Type":"ContainerStarted","Data":"b5edce904075931fac0312d5af37b9eab1b0d807ec53620d74ed700f487d867c"} Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.651349 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-f48ww\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.651419 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-config-data\") pod \"nova-cell0-conductor-db-sync-f48ww\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.651494 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-scripts\") pod \"nova-cell0-conductor-db-sync-f48ww\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.651568 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfxz4\" (UniqueName: \"kubernetes.io/projected/26734dfd-6dba-4c5f-99fc-bc27be1b8792-kube-api-access-lfxz4\") pod \"nova-cell0-conductor-db-sync-f48ww\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.659770 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-f48ww\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.663457 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-scripts\") pod \"nova-cell0-conductor-db-sync-f48ww\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.663663 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-config-data\") pod \"nova-cell0-conductor-db-sync-f48ww\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:57:56 crc kubenswrapper[4884]: W1202 01:57:56.667549 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd09bd0e7_6fdb_45f3_a2a9_c20fbc1b386c.slice/crio-ef6a9aaa0cf29dc63a91b52c8f06d49a4188a7edb8baab9d44f0cd78e430ae3b WatchSource:0}: Error finding container ef6a9aaa0cf29dc63a91b52c8f06d49a4188a7edb8baab9d44f0cd78e430ae3b: Status 404 returned error can't find the container with id ef6a9aaa0cf29dc63a91b52c8f06d49a4188a7edb8baab9d44f0cd78e430ae3b Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.673273 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfxz4\" (UniqueName: \"kubernetes.io/projected/26734dfd-6dba-4c5f-99fc-bc27be1b8792-kube-api-access-lfxz4\") pod \"nova-cell0-conductor-db-sync-f48ww\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.681220 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.689308 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.68928902 podStartE2EDuration="3.68928902s" podCreationTimestamp="2025-12-02 01:57:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:57:56.654889421 +0000 UTC m=+1293.330726315" watchObservedRunningTime="2025-12-02 01:57:56.68928902 +0000 UTC m=+1293.365125904" Dec 02 01:57:56 crc kubenswrapper[4884]: I1202 01:57:56.760332 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:57:57 crc kubenswrapper[4884]: I1202 01:57:57.276196 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-f48ww"] Dec 02 01:57:57 crc kubenswrapper[4884]: I1202 01:57:57.640064 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fdad2f7-f276-458e-96dd-6cc7fda79cac" path="/var/lib/kubelet/pods/5fdad2f7-f276-458e-96dd-6cc7fda79cac/volumes" Dec 02 01:57:57 crc kubenswrapper[4884]: I1202 01:57:57.693794 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f07d30e-bd91-4bff-bcc2-839db028de27","Type":"ContainerStarted","Data":"afaae745adae4bb340cb3fd5dda9cad148874d757560b1b5f18f3c8c77ead985"} Dec 02 01:57:57 crc kubenswrapper[4884]: I1202 01:57:57.695505 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c","Type":"ContainerStarted","Data":"d45621e05c3735377333ee659e0f348ddfafac09280377a48843037c6b344249"} Dec 02 01:57:57 crc kubenswrapper[4884]: I1202 01:57:57.695536 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c","Type":"ContainerStarted","Data":"ef6a9aaa0cf29dc63a91b52c8f06d49a4188a7edb8baab9d44f0cd78e430ae3b"} Dec 02 01:57:57 crc kubenswrapper[4884]: I1202 01:57:57.697896 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-f48ww" event={"ID":"26734dfd-6dba-4c5f-99fc-bc27be1b8792","Type":"ContainerStarted","Data":"b039bfd87e978acf2fa8017b53bda64faff015bdae9af27716625a72dede9034"} Dec 02 01:57:58 crc kubenswrapper[4884]: I1202 01:57:58.713025 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f07d30e-bd91-4bff-bcc2-839db028de27","Type":"ContainerStarted","Data":"07f68653a05b4a855702dc01743b8cb568a6c8ec80d8ce13f57302acfcf1cad4"} Dec 02 01:57:58 crc kubenswrapper[4884]: I1202 01:57:58.715707 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c","Type":"ContainerStarted","Data":"3f110cd21bbad565690372e18b78bb735074704d7a3e1cb74b1e9556844b1931"} Dec 02 01:57:58 crc kubenswrapper[4884]: I1202 01:57:58.741324 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.741304603 podStartE2EDuration="3.741304603s" podCreationTimestamp="2025-12-02 01:57:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:57:58.737009117 +0000 UTC m=+1295.412846001" watchObservedRunningTime="2025-12-02 01:57:58.741304603 +0000 UTC m=+1295.417141477" Dec 02 01:57:59 crc kubenswrapper[4884]: I1202 01:57:59.750215 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/watcher-decision-engine-0" Dec 02 01:57:59 crc kubenswrapper[4884]: I1202 01:57:59.792851 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/watcher-decision-engine-0" Dec 02 01:58:00 crc kubenswrapper[4884]: I1202 01:58:00.144761 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:58:00 crc kubenswrapper[4884]: I1202 01:58:00.745202 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f07d30e-bd91-4bff-bcc2-839db028de27","Type":"ContainerStarted","Data":"c60da44308cb4f7a8e8df6ff6f3adb26032f1b3a12ed6ab15ab090e06f521bef"} Dec 02 01:58:00 crc kubenswrapper[4884]: I1202 01:58:00.745577 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/watcher-decision-engine-0" Dec 02 01:58:00 crc kubenswrapper[4884]: I1202 01:58:00.745597 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 01:58:00 crc kubenswrapper[4884]: I1202 01:58:00.776078 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.623976948 podStartE2EDuration="6.776063209s" podCreationTimestamp="2025-12-02 01:57:54 +0000 UTC" firstStartedPulling="2025-12-02 01:57:55.450913567 +0000 UTC m=+1292.126750451" lastFinishedPulling="2025-12-02 01:57:59.602999828 +0000 UTC m=+1296.278836712" observedRunningTime="2025-12-02 01:58:00.769362644 +0000 UTC m=+1297.445199538" watchObservedRunningTime="2025-12-02 01:58:00.776063209 +0000 UTC m=+1297.451900083" Dec 02 01:58:00 crc kubenswrapper[4884]: I1202 01:58:00.779688 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/watcher-decision-engine-0" Dec 02 01:58:01 crc kubenswrapper[4884]: I1202 01:58:01.750880 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="ceilometer-central-agent" containerID="cri-o://b5edce904075931fac0312d5af37b9eab1b0d807ec53620d74ed700f487d867c" gracePeriod=30 Dec 02 01:58:01 crc kubenswrapper[4884]: I1202 01:58:01.751099 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="proxy-httpd" containerID="cri-o://c60da44308cb4f7a8e8df6ff6f3adb26032f1b3a12ed6ab15ab090e06f521bef" gracePeriod=30 Dec 02 01:58:01 crc kubenswrapper[4884]: I1202 01:58:01.751160 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="ceilometer-notification-agent" containerID="cri-o://afaae745adae4bb340cb3fd5dda9cad148874d757560b1b5f18f3c8c77ead985" gracePeriod=30 Dec 02 01:58:01 crc kubenswrapper[4884]: I1202 01:58:01.751242 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="sg-core" containerID="cri-o://07f68653a05b4a855702dc01743b8cb568a6c8ec80d8ce13f57302acfcf1cad4" gracePeriod=30 Dec 02 01:58:02 crc kubenswrapper[4884]: I1202 01:58:02.763362 4884 generic.go:334] "Generic (PLEG): container finished" podID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerID="c60da44308cb4f7a8e8df6ff6f3adb26032f1b3a12ed6ab15ab090e06f521bef" exitCode=0 Dec 02 01:58:02 crc kubenswrapper[4884]: I1202 01:58:02.763654 4884 generic.go:334] "Generic (PLEG): container finished" podID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerID="07f68653a05b4a855702dc01743b8cb568a6c8ec80d8ce13f57302acfcf1cad4" exitCode=2 Dec 02 01:58:02 crc kubenswrapper[4884]: I1202 01:58:02.763441 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f07d30e-bd91-4bff-bcc2-839db028de27","Type":"ContainerDied","Data":"c60da44308cb4f7a8e8df6ff6f3adb26032f1b3a12ed6ab15ab090e06f521bef"} Dec 02 01:58:02 crc kubenswrapper[4884]: I1202 01:58:02.763699 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f07d30e-bd91-4bff-bcc2-839db028de27","Type":"ContainerDied","Data":"07f68653a05b4a855702dc01743b8cb568a6c8ec80d8ce13f57302acfcf1cad4"} Dec 02 01:58:02 crc kubenswrapper[4884]: I1202 01:58:02.763713 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f07d30e-bd91-4bff-bcc2-839db028de27","Type":"ContainerDied","Data":"afaae745adae4bb340cb3fd5dda9cad148874d757560b1b5f18f3c8c77ead985"} Dec 02 01:58:02 crc kubenswrapper[4884]: I1202 01:58:02.763666 4884 generic.go:334] "Generic (PLEG): container finished" podID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerID="afaae745adae4bb340cb3fd5dda9cad148874d757560b1b5f18f3c8c77ead985" exitCode=0 Dec 02 01:58:03 crc kubenswrapper[4884]: I1202 01:58:03.781098 4884 generic.go:334] "Generic (PLEG): container finished" podID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerID="b5edce904075931fac0312d5af37b9eab1b0d807ec53620d74ed700f487d867c" exitCode=0 Dec 02 01:58:03 crc kubenswrapper[4884]: I1202 01:58:03.781318 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f07d30e-bd91-4bff-bcc2-839db028de27","Type":"ContainerDied","Data":"b5edce904075931fac0312d5af37b9eab1b0d807ec53620d74ed700f487d867c"} Dec 02 01:58:03 crc kubenswrapper[4884]: I1202 01:58:03.971941 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 02 01:58:03 crc kubenswrapper[4884]: I1202 01:58:03.973128 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 02 01:58:04 crc kubenswrapper[4884]: I1202 01:58:04.007913 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 02 01:58:04 crc kubenswrapper[4884]: I1202 01:58:04.018550 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 02 01:58:04 crc kubenswrapper[4884]: I1202 01:58:04.792302 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 02 01:58:04 crc kubenswrapper[4884]: I1202 01:58:04.792345 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 02 01:58:05 crc kubenswrapper[4884]: I1202 01:58:05.928633 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.035269 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-config-data\") pod \"7f07d30e-bd91-4bff-bcc2-839db028de27\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.035316 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tczdp\" (UniqueName: \"kubernetes.io/projected/7f07d30e-bd91-4bff-bcc2-839db028de27-kube-api-access-tczdp\") pod \"7f07d30e-bd91-4bff-bcc2-839db028de27\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.035360 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-scripts\") pod \"7f07d30e-bd91-4bff-bcc2-839db028de27\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.035394 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-combined-ca-bundle\") pod \"7f07d30e-bd91-4bff-bcc2-839db028de27\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.035469 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f07d30e-bd91-4bff-bcc2-839db028de27-log-httpd\") pod \"7f07d30e-bd91-4bff-bcc2-839db028de27\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.035546 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-sg-core-conf-yaml\") pod \"7f07d30e-bd91-4bff-bcc2-839db028de27\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.035635 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f07d30e-bd91-4bff-bcc2-839db028de27-run-httpd\") pod \"7f07d30e-bd91-4bff-bcc2-839db028de27\" (UID: \"7f07d30e-bd91-4bff-bcc2-839db028de27\") " Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.036347 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f07d30e-bd91-4bff-bcc2-839db028de27-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7f07d30e-bd91-4bff-bcc2-839db028de27" (UID: "7f07d30e-bd91-4bff-bcc2-839db028de27"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.037316 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f07d30e-bd91-4bff-bcc2-839db028de27-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7f07d30e-bd91-4bff-bcc2-839db028de27" (UID: "7f07d30e-bd91-4bff-bcc2-839db028de27"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.047733 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-scripts" (OuterVolumeSpecName: "scripts") pod "7f07d30e-bd91-4bff-bcc2-839db028de27" (UID: "7f07d30e-bd91-4bff-bcc2-839db028de27"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.047929 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f07d30e-bd91-4bff-bcc2-839db028de27-kube-api-access-tczdp" (OuterVolumeSpecName: "kube-api-access-tczdp") pod "7f07d30e-bd91-4bff-bcc2-839db028de27" (UID: "7f07d30e-bd91-4bff-bcc2-839db028de27"). InnerVolumeSpecName "kube-api-access-tczdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.072152 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7f07d30e-bd91-4bff-bcc2-839db028de27" (UID: "7f07d30e-bd91-4bff-bcc2-839db028de27"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.094864 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.094915 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.122605 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f07d30e-bd91-4bff-bcc2-839db028de27" (UID: "7f07d30e-bd91-4bff-bcc2-839db028de27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.131349 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.139383 4884 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.139419 4884 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f07d30e-bd91-4bff-bcc2-839db028de27-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.139427 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tczdp\" (UniqueName: \"kubernetes.io/projected/7f07d30e-bd91-4bff-bcc2-839db028de27-kube-api-access-tczdp\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.139438 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.139446 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.139455 4884 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7f07d30e-bd91-4bff-bcc2-839db028de27-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.143309 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-config-data" (OuterVolumeSpecName: "config-data") pod "7f07d30e-bd91-4bff-bcc2-839db028de27" (UID: "7f07d30e-bd91-4bff-bcc2-839db028de27"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.147701 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.241794 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f07d30e-bd91-4bff-bcc2-839db028de27-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.621736 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.723652 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.823806 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-f48ww" event={"ID":"26734dfd-6dba-4c5f-99fc-bc27be1b8792","Type":"ContainerStarted","Data":"77e7351dbbd5eabe7d0c4f22fd6464f261cfa31d100e6317bc505e1d5cfa2135"} Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.836127 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.836639 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7f07d30e-bd91-4bff-bcc2-839db028de27","Type":"ContainerDied","Data":"a159f7916573de4e3d20ba4dd94bd17d3fcbd46c36865c29be69dd1ea38fb465"} Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.836688 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.836710 4884 scope.go:117] "RemoveContainer" containerID="c60da44308cb4f7a8e8df6ff6f3adb26032f1b3a12ed6ab15ab090e06f521bef" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.836995 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.851835 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-f48ww" podStartSLOduration=2.484001602 podStartE2EDuration="10.851816748s" podCreationTimestamp="2025-12-02 01:57:56 +0000 UTC" firstStartedPulling="2025-12-02 01:57:57.276608001 +0000 UTC m=+1293.952444885" lastFinishedPulling="2025-12-02 01:58:05.644423147 +0000 UTC m=+1302.320260031" observedRunningTime="2025-12-02 01:58:06.846096107 +0000 UTC m=+1303.521932991" watchObservedRunningTime="2025-12-02 01:58:06.851816748 +0000 UTC m=+1303.527653642" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.878283 4884 scope.go:117] "RemoveContainer" containerID="07f68653a05b4a855702dc01743b8cb568a6c8ec80d8ce13f57302acfcf1cad4" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.913361 4884 scope.go:117] "RemoveContainer" containerID="afaae745adae4bb340cb3fd5dda9cad148874d757560b1b5f18f3c8c77ead985" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.941854 4884 scope.go:117] "RemoveContainer" containerID="b5edce904075931fac0312d5af37b9eab1b0d807ec53620d74ed700f487d867c" Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.944000 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:58:06 crc kubenswrapper[4884]: I1202 01:58:06.959995 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.019977 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:58:07 crc kubenswrapper[4884]: E1202 01:58:07.020497 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="ceilometer-notification-agent" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.020515 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="ceilometer-notification-agent" Dec 02 01:58:07 crc kubenswrapper[4884]: E1202 01:58:07.020532 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="sg-core" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.020540 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="sg-core" Dec 02 01:58:07 crc kubenswrapper[4884]: E1202 01:58:07.020551 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="ceilometer-central-agent" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.020561 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="ceilometer-central-agent" Dec 02 01:58:07 crc kubenswrapper[4884]: E1202 01:58:07.020596 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="proxy-httpd" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.020604 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="proxy-httpd" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.020850 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="proxy-httpd" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.020873 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="ceilometer-central-agent" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.020886 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="sg-core" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.020906 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" containerName="ceilometer-notification-agent" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.023113 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.028713 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.029080 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.036625 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.074403 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.074657 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.074804 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6e289da-5ac9-4191-8bf1-2611786733d5-log-httpd\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.074930 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-config-data\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.075041 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj7kg\" (UniqueName: \"kubernetes.io/projected/d6e289da-5ac9-4191-8bf1-2611786733d5-kube-api-access-qj7kg\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.075196 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6e289da-5ac9-4191-8bf1-2611786733d5-run-httpd\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.075304 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-scripts\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.176783 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.176835 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6e289da-5ac9-4191-8bf1-2611786733d5-log-httpd\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.176858 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-config-data\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.176904 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj7kg\" (UniqueName: \"kubernetes.io/projected/d6e289da-5ac9-4191-8bf1-2611786733d5-kube-api-access-qj7kg\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.176977 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6e289da-5ac9-4191-8bf1-2611786733d5-run-httpd\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.177014 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-scripts\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.177041 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.177451 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6e289da-5ac9-4191-8bf1-2611786733d5-log-httpd\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.177690 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6e289da-5ac9-4191-8bf1-2611786733d5-run-httpd\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.181611 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.181879 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-scripts\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.182176 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-config-data\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.194191 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.195057 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj7kg\" (UniqueName: \"kubernetes.io/projected/d6e289da-5ac9-4191-8bf1-2611786733d5-kube-api-access-qj7kg\") pod \"ceilometer-0\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.363619 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.629414 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f07d30e-bd91-4bff-bcc2-839db028de27" path="/var/lib/kubelet/pods/7f07d30e-bd91-4bff-bcc2-839db028de27/volumes" Dec 02 01:58:07 crc kubenswrapper[4884]: I1202 01:58:07.859484 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:58:07 crc kubenswrapper[4884]: W1202 01:58:07.880048 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6e289da_5ac9_4191_8bf1_2611786733d5.slice/crio-7d7dd89f0fde1dad0ed47c5c6271f9adc44dc420a7f156c81942e28406bbeee2 WatchSource:0}: Error finding container 7d7dd89f0fde1dad0ed47c5c6271f9adc44dc420a7f156c81942e28406bbeee2: Status 404 returned error can't find the container with id 7d7dd89f0fde1dad0ed47c5c6271f9adc44dc420a7f156c81942e28406bbeee2 Dec 02 01:58:08 crc kubenswrapper[4884]: I1202 01:58:08.770140 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 02 01:58:08 crc kubenswrapper[4884]: I1202 01:58:08.864600 4884 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 02 01:58:08 crc kubenswrapper[4884]: I1202 01:58:08.865682 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6e289da-5ac9-4191-8bf1-2611786733d5","Type":"ContainerStarted","Data":"2909efeea26d7f14bc4d65944e00b64dceb7904c66064476b993decf02b781db"} Dec 02 01:58:08 crc kubenswrapper[4884]: I1202 01:58:08.865725 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6e289da-5ac9-4191-8bf1-2611786733d5","Type":"ContainerStarted","Data":"7d7dd89f0fde1dad0ed47c5c6271f9adc44dc420a7f156c81942e28406bbeee2"} Dec 02 01:58:09 crc kubenswrapper[4884]: I1202 01:58:09.211105 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 02 01:58:09 crc kubenswrapper[4884]: I1202 01:58:09.890416 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6e289da-5ac9-4191-8bf1-2611786733d5","Type":"ContainerStarted","Data":"c9e19a9dfbb3d9d2019696ffd26fdfa8e5652b4d1ba8ce862f6be658dae9baf9"} Dec 02 01:58:10 crc kubenswrapper[4884]: I1202 01:58:10.901992 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6e289da-5ac9-4191-8bf1-2611786733d5","Type":"ContainerStarted","Data":"a026cb2e4e0d8ebc6078faa721738be77d0dabd0ba1f95c630b264b1b84a471f"} Dec 02 01:58:11 crc kubenswrapper[4884]: I1202 01:58:11.913247 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6e289da-5ac9-4191-8bf1-2611786733d5","Type":"ContainerStarted","Data":"e822efae88553c322dcf0063bcb0396d1dd3f2a9969a1751f69a70177c6a82e6"} Dec 02 01:58:11 crc kubenswrapper[4884]: I1202 01:58:11.913583 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 01:58:11 crc kubenswrapper[4884]: I1202 01:58:11.946572 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.452998198 podStartE2EDuration="5.946551931s" podCreationTimestamp="2025-12-02 01:58:06 +0000 UTC" firstStartedPulling="2025-12-02 01:58:07.882763052 +0000 UTC m=+1304.558599936" lastFinishedPulling="2025-12-02 01:58:11.376316745 +0000 UTC m=+1308.052153669" observedRunningTime="2025-12-02 01:58:11.936650647 +0000 UTC m=+1308.612487551" watchObservedRunningTime="2025-12-02 01:58:11.946551931 +0000 UTC m=+1308.622388815" Dec 02 01:58:18 crc kubenswrapper[4884]: I1202 01:58:18.992571 4884 generic.go:334] "Generic (PLEG): container finished" podID="26734dfd-6dba-4c5f-99fc-bc27be1b8792" containerID="77e7351dbbd5eabe7d0c4f22fd6464f261cfa31d100e6317bc505e1d5cfa2135" exitCode=0 Dec 02 01:58:18 crc kubenswrapper[4884]: I1202 01:58:18.992669 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-f48ww" event={"ID":"26734dfd-6dba-4c5f-99fc-bc27be1b8792","Type":"ContainerDied","Data":"77e7351dbbd5eabe7d0c4f22fd6464f261cfa31d100e6317bc505e1d5cfa2135"} Dec 02 01:58:20 crc kubenswrapper[4884]: I1202 01:58:20.404840 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:58:20 crc kubenswrapper[4884]: I1202 01:58:20.484496 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-scripts\") pod \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " Dec 02 01:58:20 crc kubenswrapper[4884]: I1202 01:58:20.484554 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfxz4\" (UniqueName: \"kubernetes.io/projected/26734dfd-6dba-4c5f-99fc-bc27be1b8792-kube-api-access-lfxz4\") pod \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " Dec 02 01:58:20 crc kubenswrapper[4884]: I1202 01:58:20.484696 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-config-data\") pod \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " Dec 02 01:58:20 crc kubenswrapper[4884]: I1202 01:58:20.484864 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-combined-ca-bundle\") pod \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\" (UID: \"26734dfd-6dba-4c5f-99fc-bc27be1b8792\") " Dec 02 01:58:20 crc kubenswrapper[4884]: I1202 01:58:20.490854 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26734dfd-6dba-4c5f-99fc-bc27be1b8792-kube-api-access-lfxz4" (OuterVolumeSpecName: "kube-api-access-lfxz4") pod "26734dfd-6dba-4c5f-99fc-bc27be1b8792" (UID: "26734dfd-6dba-4c5f-99fc-bc27be1b8792"). InnerVolumeSpecName "kube-api-access-lfxz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:58:20 crc kubenswrapper[4884]: I1202 01:58:20.494582 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-scripts" (OuterVolumeSpecName: "scripts") pod "26734dfd-6dba-4c5f-99fc-bc27be1b8792" (UID: "26734dfd-6dba-4c5f-99fc-bc27be1b8792"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:20 crc kubenswrapper[4884]: I1202 01:58:20.521906 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-config-data" (OuterVolumeSpecName: "config-data") pod "26734dfd-6dba-4c5f-99fc-bc27be1b8792" (UID: "26734dfd-6dba-4c5f-99fc-bc27be1b8792"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:20 crc kubenswrapper[4884]: I1202 01:58:20.540121 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26734dfd-6dba-4c5f-99fc-bc27be1b8792" (UID: "26734dfd-6dba-4c5f-99fc-bc27be1b8792"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:20 crc kubenswrapper[4884]: I1202 01:58:20.587599 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:20 crc kubenswrapper[4884]: I1202 01:58:20.587638 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfxz4\" (UniqueName: \"kubernetes.io/projected/26734dfd-6dba-4c5f-99fc-bc27be1b8792-kube-api-access-lfxz4\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:20 crc kubenswrapper[4884]: I1202 01:58:20.587651 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:20 crc kubenswrapper[4884]: I1202 01:58:20.587664 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26734dfd-6dba-4c5f-99fc-bc27be1b8792-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.020580 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-f48ww" event={"ID":"26734dfd-6dba-4c5f-99fc-bc27be1b8792","Type":"ContainerDied","Data":"b039bfd87e978acf2fa8017b53bda64faff015bdae9af27716625a72dede9034"} Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.020620 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b039bfd87e978acf2fa8017b53bda64faff015bdae9af27716625a72dede9034" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.020667 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-f48ww" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.213370 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 02 01:58:21 crc kubenswrapper[4884]: E1202 01:58:21.214258 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26734dfd-6dba-4c5f-99fc-bc27be1b8792" containerName="nova-cell0-conductor-db-sync" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.214285 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="26734dfd-6dba-4c5f-99fc-bc27be1b8792" containerName="nova-cell0-conductor-db-sync" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.214596 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="26734dfd-6dba-4c5f-99fc-bc27be1b8792" containerName="nova-cell0-conductor-db-sync" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.215849 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.219295 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-52l2h" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.221886 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.226505 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.298603 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d11b36-d993-46f4-9ca9-813df373dcab-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"67d11b36-d993-46f4-9ca9-813df373dcab\") " pod="openstack/nova-cell0-conductor-0" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.298691 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d11b36-d993-46f4-9ca9-813df373dcab-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"67d11b36-d993-46f4-9ca9-813df373dcab\") " pod="openstack/nova-cell0-conductor-0" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.298839 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7xbg\" (UniqueName: \"kubernetes.io/projected/67d11b36-d993-46f4-9ca9-813df373dcab-kube-api-access-v7xbg\") pod \"nova-cell0-conductor-0\" (UID: \"67d11b36-d993-46f4-9ca9-813df373dcab\") " pod="openstack/nova-cell0-conductor-0" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.400550 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d11b36-d993-46f4-9ca9-813df373dcab-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"67d11b36-d993-46f4-9ca9-813df373dcab\") " pod="openstack/nova-cell0-conductor-0" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.400642 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d11b36-d993-46f4-9ca9-813df373dcab-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"67d11b36-d993-46f4-9ca9-813df373dcab\") " pod="openstack/nova-cell0-conductor-0" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.400728 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7xbg\" (UniqueName: \"kubernetes.io/projected/67d11b36-d993-46f4-9ca9-813df373dcab-kube-api-access-v7xbg\") pod \"nova-cell0-conductor-0\" (UID: \"67d11b36-d993-46f4-9ca9-813df373dcab\") " pod="openstack/nova-cell0-conductor-0" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.405818 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67d11b36-d993-46f4-9ca9-813df373dcab-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"67d11b36-d993-46f4-9ca9-813df373dcab\") " pod="openstack/nova-cell0-conductor-0" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.406442 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67d11b36-d993-46f4-9ca9-813df373dcab-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"67d11b36-d993-46f4-9ca9-813df373dcab\") " pod="openstack/nova-cell0-conductor-0" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.417396 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7xbg\" (UniqueName: \"kubernetes.io/projected/67d11b36-d993-46f4-9ca9-813df373dcab-kube-api-access-v7xbg\") pod \"nova-cell0-conductor-0\" (UID: \"67d11b36-d993-46f4-9ca9-813df373dcab\") " pod="openstack/nova-cell0-conductor-0" Dec 02 01:58:21 crc kubenswrapper[4884]: I1202 01:58:21.620830 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 02 01:58:22 crc kubenswrapper[4884]: I1202 01:58:22.121467 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 02 01:58:22 crc kubenswrapper[4884]: W1202 01:58:22.126100 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67d11b36_d993_46f4_9ca9_813df373dcab.slice/crio-277d91f01d2893bb1d1e06883ee8bff9959d9ebabc8c39cf315cf9e4ded9d59a WatchSource:0}: Error finding container 277d91f01d2893bb1d1e06883ee8bff9959d9ebabc8c39cf315cf9e4ded9d59a: Status 404 returned error can't find the container with id 277d91f01d2893bb1d1e06883ee8bff9959d9ebabc8c39cf315cf9e4ded9d59a Dec 02 01:58:23 crc kubenswrapper[4884]: I1202 01:58:23.037638 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"67d11b36-d993-46f4-9ca9-813df373dcab","Type":"ContainerStarted","Data":"177ec0cb0feb45aabaaa659b9f89ef4f023ad8cd26711f3973a554e93f19fd7e"} Dec 02 01:58:23 crc kubenswrapper[4884]: I1202 01:58:23.037921 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"67d11b36-d993-46f4-9ca9-813df373dcab","Type":"ContainerStarted","Data":"277d91f01d2893bb1d1e06883ee8bff9959d9ebabc8c39cf315cf9e4ded9d59a"} Dec 02 01:58:23 crc kubenswrapper[4884]: I1202 01:58:23.037961 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 02 01:58:31 crc kubenswrapper[4884]: I1202 01:58:31.654101 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 02 01:58:31 crc kubenswrapper[4884]: I1202 01:58:31.683855 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=10.68383609 podStartE2EDuration="10.68383609s" podCreationTimestamp="2025-12-02 01:58:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:58:23.05807663 +0000 UTC m=+1319.733913524" watchObservedRunningTime="2025-12-02 01:58:31.68383609 +0000 UTC m=+1328.359672974" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.183447 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-x695l"] Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.185010 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.188095 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.193861 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.194368 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-x695l"] Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.344026 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.345621 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.347383 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.347883 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsblp\" (UniqueName: \"kubernetes.io/projected/5a0a5720-88c1-4adc-9f54-2db975a9e48a-kube-api-access-wsblp\") pod \"nova-cell0-cell-mapping-x695l\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.348054 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-config-data\") pod \"nova-cell0-cell-mapping-x695l\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.348166 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x695l\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.348212 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-scripts\") pod \"nova-cell0-cell-mapping-x695l\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.356420 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.400883 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.402205 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.403941 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.413780 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.450910 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-config-data\") pod \"nova-cell0-cell-mapping-x695l\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.450968 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69686a6c-cdef-4fdf-8838-2624f841a85a-config-data\") pod \"nova-api-0\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " pod="openstack/nova-api-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.451026 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x695l\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.451054 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-scripts\") pod \"nova-cell0-cell-mapping-x695l\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.451083 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tk6d\" (UniqueName: \"kubernetes.io/projected/69686a6c-cdef-4fdf-8838-2624f841a85a-kube-api-access-8tk6d\") pod \"nova-api-0\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " pod="openstack/nova-api-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.451099 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsblp\" (UniqueName: \"kubernetes.io/projected/5a0a5720-88c1-4adc-9f54-2db975a9e48a-kube-api-access-wsblp\") pod \"nova-cell0-cell-mapping-x695l\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.451157 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69686a6c-cdef-4fdf-8838-2624f841a85a-logs\") pod \"nova-api-0\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " pod="openstack/nova-api-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.451172 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69686a6c-cdef-4fdf-8838-2624f841a85a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " pod="openstack/nova-api-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.463590 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-config-data\") pod \"nova-cell0-cell-mapping-x695l\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.465303 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x695l\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.466189 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-scripts\") pod \"nova-cell0-cell-mapping-x695l\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.474638 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsblp\" (UniqueName: \"kubernetes.io/projected/5a0a5720-88c1-4adc-9f54-2db975a9e48a-kube-api-access-wsblp\") pod \"nova-cell0-cell-mapping-x695l\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.507231 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.551823 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.553022 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.555173 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69686a6c-cdef-4fdf-8838-2624f841a85a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " pod="openstack/nova-api-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.555202 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69686a6c-cdef-4fdf-8838-2624f841a85a-logs\") pod \"nova-api-0\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " pod="openstack/nova-api-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.555251 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98757586-4157-403c-8a34-f7d28f2451cc-config-data\") pod \"nova-scheduler-0\" (UID: \"98757586-4157-403c-8a34-f7d28f2451cc\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.555286 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69686a6c-cdef-4fdf-8838-2624f841a85a-config-data\") pod \"nova-api-0\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " pod="openstack/nova-api-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.555315 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzvf4\" (UniqueName: \"kubernetes.io/projected/98757586-4157-403c-8a34-f7d28f2451cc-kube-api-access-rzvf4\") pod \"nova-scheduler-0\" (UID: \"98757586-4157-403c-8a34-f7d28f2451cc\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.555373 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tk6d\" (UniqueName: \"kubernetes.io/projected/69686a6c-cdef-4fdf-8838-2624f841a85a-kube-api-access-8tk6d\") pod \"nova-api-0\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " pod="openstack/nova-api-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.555406 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98757586-4157-403c-8a34-f7d28f2451cc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"98757586-4157-403c-8a34-f7d28f2451cc\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.562367 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.562728 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69686a6c-cdef-4fdf-8838-2624f841a85a-logs\") pod \"nova-api-0\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " pod="openstack/nova-api-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.563681 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69686a6c-cdef-4fdf-8838-2624f841a85a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " pod="openstack/nova-api-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.578825 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.580405 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.588541 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69686a6c-cdef-4fdf-8838-2624f841a85a-config-data\") pod \"nova-api-0\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " pod="openstack/nova-api-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.589324 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.603557 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tk6d\" (UniqueName: \"kubernetes.io/projected/69686a6c-cdef-4fdf-8838-2624f841a85a-kube-api-access-8tk6d\") pod \"nova-api-0\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " pod="openstack/nova-api-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.625928 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.656874 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98757586-4157-403c-8a34-f7d28f2451cc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"98757586-4157-403c-8a34-f7d28f2451cc\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.656971 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98757586-4157-403c-8a34-f7d28f2451cc-config-data\") pod \"nova-scheduler-0\" (UID: \"98757586-4157-403c-8a34-f7d28f2451cc\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.657015 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzvf4\" (UniqueName: \"kubernetes.io/projected/98757586-4157-403c-8a34-f7d28f2451cc-kube-api-access-rzvf4\") pod \"nova-scheduler-0\" (UID: \"98757586-4157-403c-8a34-f7d28f2451cc\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.657046 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5wsq\" (UniqueName: \"kubernetes.io/projected/bf6e71cc-1f65-462c-8234-3c93203d1728-kube-api-access-f5wsq\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf6e71cc-1f65-462c-8234-3c93203d1728\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.657085 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf6e71cc-1f65-462c-8234-3c93203d1728-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf6e71cc-1f65-462c-8234-3c93203d1728\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.657110 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf6e71cc-1f65-462c-8234-3c93203d1728-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf6e71cc-1f65-462c-8234-3c93203d1728\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.661563 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98757586-4157-403c-8a34-f7d28f2451cc-config-data\") pod \"nova-scheduler-0\" (UID: \"98757586-4157-403c-8a34-f7d28f2451cc\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.665018 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98757586-4157-403c-8a34-f7d28f2451cc-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"98757586-4157-403c-8a34-f7d28f2451cc\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.665167 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.680166 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.691337 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzvf4\" (UniqueName: \"kubernetes.io/projected/98757586-4157-403c-8a34-f7d28f2451cc-kube-api-access-rzvf4\") pod \"nova-scheduler-0\" (UID: \"98757586-4157-403c-8a34-f7d28f2451cc\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.720585 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.728356 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-b74v6"] Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.730333 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.758144 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-b74v6"] Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.759877 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d4c86db-e57a-4e79-a242-fe623b71f59b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " pod="openstack/nova-metadata-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.760023 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5wsq\" (UniqueName: \"kubernetes.io/projected/bf6e71cc-1f65-462c-8234-3c93203d1728-kube-api-access-f5wsq\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf6e71cc-1f65-462c-8234-3c93203d1728\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.760102 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d4c86db-e57a-4e79-a242-fe623b71f59b-logs\") pod \"nova-metadata-0\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " pod="openstack/nova-metadata-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.760135 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf6e71cc-1f65-462c-8234-3c93203d1728-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf6e71cc-1f65-462c-8234-3c93203d1728\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.760161 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf6e71cc-1f65-462c-8234-3c93203d1728-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf6e71cc-1f65-462c-8234-3c93203d1728\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.760212 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d4c86db-e57a-4e79-a242-fe623b71f59b-config-data\") pod \"nova-metadata-0\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " pod="openstack/nova-metadata-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.760242 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jtv5\" (UniqueName: \"kubernetes.io/projected/5d4c86db-e57a-4e79-a242-fe623b71f59b-kube-api-access-8jtv5\") pod \"nova-metadata-0\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " pod="openstack/nova-metadata-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.770851 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf6e71cc-1f65-462c-8234-3c93203d1728-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf6e71cc-1f65-462c-8234-3c93203d1728\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.782223 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf6e71cc-1f65-462c-8234-3c93203d1728-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf6e71cc-1f65-462c-8234-3c93203d1728\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.786863 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5wsq\" (UniqueName: \"kubernetes.io/projected/bf6e71cc-1f65-462c-8234-3c93203d1728-kube-api-access-f5wsq\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf6e71cc-1f65-462c-8234-3c93203d1728\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.862248 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d4c86db-e57a-4e79-a242-fe623b71f59b-config-data\") pod \"nova-metadata-0\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " pod="openstack/nova-metadata-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.862398 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jtv5\" (UniqueName: \"kubernetes.io/projected/5d4c86db-e57a-4e79-a242-fe623b71f59b-kube-api-access-8jtv5\") pod \"nova-metadata-0\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " pod="openstack/nova-metadata-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.862448 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.862509 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-dns-svc\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.862562 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-config\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.862597 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.862641 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d4c86db-e57a-4e79-a242-fe623b71f59b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " pod="openstack/nova-metadata-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.862735 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d4c86db-e57a-4e79-a242-fe623b71f59b-logs\") pod \"nova-metadata-0\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " pod="openstack/nova-metadata-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.862875 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw9bv\" (UniqueName: \"kubernetes.io/projected/2686aabc-d1de-4d56-95db-0b919d055b3d-kube-api-access-sw9bv\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.863032 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.868203 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d4c86db-e57a-4e79-a242-fe623b71f59b-config-data\") pod \"nova-metadata-0\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " pod="openstack/nova-metadata-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.869781 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d4c86db-e57a-4e79-a242-fe623b71f59b-logs\") pod \"nova-metadata-0\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " pod="openstack/nova-metadata-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.881799 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d4c86db-e57a-4e79-a242-fe623b71f59b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " pod="openstack/nova-metadata-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.907847 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jtv5\" (UniqueName: \"kubernetes.io/projected/5d4c86db-e57a-4e79-a242-fe623b71f59b-kube-api-access-8jtv5\") pod \"nova-metadata-0\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " pod="openstack/nova-metadata-0" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.964902 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-config\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.964957 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.965062 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw9bv\" (UniqueName: \"kubernetes.io/projected/2686aabc-d1de-4d56-95db-0b919d055b3d-kube-api-access-sw9bv\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.965113 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.965186 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.965237 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-dns-svc\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.966236 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-dns-svc\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.966895 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-config\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.967493 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.967622 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.968367 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:32 crc kubenswrapper[4884]: I1202 01:58:32.998172 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw9bv\" (UniqueName: \"kubernetes.io/projected/2686aabc-d1de-4d56-95db-0b919d055b3d-kube-api-access-sw9bv\") pod \"dnsmasq-dns-757b4f8459-b74v6\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.001087 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.015165 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.067644 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.180888 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-x695l"] Dec 02 01:58:33 crc kubenswrapper[4884]: W1202 01:58:33.214156 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a0a5720_88c1_4adc_9f54_2db975a9e48a.slice/crio-c09c82e60c9ede7ff70afa2ba6993db2eeab484eca31542a4bfd8bffd93c974d WatchSource:0}: Error finding container c09c82e60c9ede7ff70afa2ba6993db2eeab484eca31542a4bfd8bffd93c974d: Status 404 returned error can't find the container with id c09c82e60c9ede7ff70afa2ba6993db2eeab484eca31542a4bfd8bffd93c974d Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.269118 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:58:33 crc kubenswrapper[4884]: W1202 01:58:33.318302 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69686a6c_cdef_4fdf_8838_2624f841a85a.slice/crio-4188b9e7dcc7cdd426348da93f02fe6386d14ae7e6331066d7b529c037e78220 WatchSource:0}: Error finding container 4188b9e7dcc7cdd426348da93f02fe6386d14ae7e6331066d7b529c037e78220: Status 404 returned error can't find the container with id 4188b9e7dcc7cdd426348da93f02fe6386d14ae7e6331066d7b529c037e78220 Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.335819 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-kcpzh"] Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.340983 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.345345 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.345533 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.356062 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-kcpzh"] Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.388001 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.395130 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdbkt\" (UniqueName: \"kubernetes.io/projected/e74647f6-5b85-47c2-9f26-27465f963095-kube-api-access-kdbkt\") pod \"nova-cell1-conductor-db-sync-kcpzh\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.395413 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-kcpzh\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.395843 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-scripts\") pod \"nova-cell1-conductor-db-sync-kcpzh\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.395925 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-config-data\") pod \"nova-cell1-conductor-db-sync-kcpzh\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.500034 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdbkt\" (UniqueName: \"kubernetes.io/projected/e74647f6-5b85-47c2-9f26-27465f963095-kube-api-access-kdbkt\") pod \"nova-cell1-conductor-db-sync-kcpzh\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.500120 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-kcpzh\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.500182 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-scripts\") pod \"nova-cell1-conductor-db-sync-kcpzh\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.500201 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-config-data\") pod \"nova-cell1-conductor-db-sync-kcpzh\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.504887 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-scripts\") pod \"nova-cell1-conductor-db-sync-kcpzh\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.508636 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-kcpzh\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.508846 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-config-data\") pod \"nova-cell1-conductor-db-sync-kcpzh\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.522159 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdbkt\" (UniqueName: \"kubernetes.io/projected/e74647f6-5b85-47c2-9f26-27465f963095-kube-api-access-kdbkt\") pod \"nova-cell1-conductor-db-sync-kcpzh\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.577994 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:58:33 crc kubenswrapper[4884]: W1202 01:58:33.583714 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d4c86db_e57a_4e79_a242_fe623b71f59b.slice/crio-b78b28cecaf9de4f9d0752bd5ba6050cdd9f157f77b053d45d84ba93fe957c96 WatchSource:0}: Error finding container b78b28cecaf9de4f9d0752bd5ba6050cdd9f157f77b053d45d84ba93fe957c96: Status 404 returned error can't find the container with id b78b28cecaf9de4f9d0752bd5ba6050cdd9f157f77b053d45d84ba93fe957c96 Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.681596 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.813829 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 01:58:33 crc kubenswrapper[4884]: I1202 01:58:33.944795 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-b74v6"] Dec 02 01:58:34 crc kubenswrapper[4884]: I1202 01:58:34.152774 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d4c86db-e57a-4e79-a242-fe623b71f59b","Type":"ContainerStarted","Data":"b78b28cecaf9de4f9d0752bd5ba6050cdd9f157f77b053d45d84ba93fe957c96"} Dec 02 01:58:34 crc kubenswrapper[4884]: I1202 01:58:34.156442 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-b74v6" event={"ID":"2686aabc-d1de-4d56-95db-0b919d055b3d","Type":"ContainerStarted","Data":"6b5585bce3361f67293c467b46b6db49399765fadfcc4deefd1a020991104bcf"} Dec 02 01:58:34 crc kubenswrapper[4884]: I1202 01:58:34.157723 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"98757586-4157-403c-8a34-f7d28f2451cc","Type":"ContainerStarted","Data":"3b7b070da3fa28f84f8934e769745bd930e042b6a030252dc58e38db24afab6d"} Dec 02 01:58:34 crc kubenswrapper[4884]: I1202 01:58:34.159422 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"69686a6c-cdef-4fdf-8838-2624f841a85a","Type":"ContainerStarted","Data":"4188b9e7dcc7cdd426348da93f02fe6386d14ae7e6331066d7b529c037e78220"} Dec 02 01:58:34 crc kubenswrapper[4884]: I1202 01:58:34.161618 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bf6e71cc-1f65-462c-8234-3c93203d1728","Type":"ContainerStarted","Data":"d5708b5383466dc627b5aed79db3e94818e0a8ba2546ffe7370f4afce88345e4"} Dec 02 01:58:34 crc kubenswrapper[4884]: I1202 01:58:34.163201 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x695l" event={"ID":"5a0a5720-88c1-4adc-9f54-2db975a9e48a","Type":"ContainerStarted","Data":"a4b47a3ed380d29bda07b1cc310be577ed09eece497e9ca981028164b29fc5a4"} Dec 02 01:58:34 crc kubenswrapper[4884]: I1202 01:58:34.163259 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x695l" event={"ID":"5a0a5720-88c1-4adc-9f54-2db975a9e48a","Type":"ContainerStarted","Data":"c09c82e60c9ede7ff70afa2ba6993db2eeab484eca31542a4bfd8bffd93c974d"} Dec 02 01:58:34 crc kubenswrapper[4884]: I1202 01:58:34.189226 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-x695l" podStartSLOduration=2.189207262 podStartE2EDuration="2.189207262s" podCreationTimestamp="2025-12-02 01:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:58:34.182048036 +0000 UTC m=+1330.857884920" watchObservedRunningTime="2025-12-02 01:58:34.189207262 +0000 UTC m=+1330.865044146" Dec 02 01:58:34 crc kubenswrapper[4884]: I1202 01:58:34.300798 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-kcpzh"] Dec 02 01:58:34 crc kubenswrapper[4884]: W1202 01:58:34.300946 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode74647f6_5b85_47c2_9f26_27465f963095.slice/crio-108530649f545fcc5e409e072f9dbafbd2a34449bfbe6ee518039bbe6cb7ac52 WatchSource:0}: Error finding container 108530649f545fcc5e409e072f9dbafbd2a34449bfbe6ee518039bbe6cb7ac52: Status 404 returned error can't find the container with id 108530649f545fcc5e409e072f9dbafbd2a34449bfbe6ee518039bbe6cb7ac52 Dec 02 01:58:35 crc kubenswrapper[4884]: I1202 01:58:35.175360 4884 generic.go:334] "Generic (PLEG): container finished" podID="2686aabc-d1de-4d56-95db-0b919d055b3d" containerID="e70c728f8da1585c72cc4e4a885c9377b7f2957a5aac621406e67efdea96a6b6" exitCode=0 Dec 02 01:58:35 crc kubenswrapper[4884]: I1202 01:58:35.175492 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-b74v6" event={"ID":"2686aabc-d1de-4d56-95db-0b919d055b3d","Type":"ContainerDied","Data":"e70c728f8da1585c72cc4e4a885c9377b7f2957a5aac621406e67efdea96a6b6"} Dec 02 01:58:35 crc kubenswrapper[4884]: I1202 01:58:35.177632 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-kcpzh" event={"ID":"e74647f6-5b85-47c2-9f26-27465f963095","Type":"ContainerStarted","Data":"e54c5930d9ab03a015536b995f4926f387f59c08dd7848b2af7b37b877f67176"} Dec 02 01:58:35 crc kubenswrapper[4884]: I1202 01:58:35.177679 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-kcpzh" event={"ID":"e74647f6-5b85-47c2-9f26-27465f963095","Type":"ContainerStarted","Data":"108530649f545fcc5e409e072f9dbafbd2a34449bfbe6ee518039bbe6cb7ac52"} Dec 02 01:58:35 crc kubenswrapper[4884]: I1202 01:58:35.221520 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-kcpzh" podStartSLOduration=2.22150369 podStartE2EDuration="2.22150369s" podCreationTimestamp="2025-12-02 01:58:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:58:35.215136873 +0000 UTC m=+1331.890973767" watchObservedRunningTime="2025-12-02 01:58:35.22150369 +0000 UTC m=+1331.897340574" Dec 02 01:58:35 crc kubenswrapper[4884]: I1202 01:58:35.789423 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:58:35 crc kubenswrapper[4884]: I1202 01:58:35.800486 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 01:58:37 crc kubenswrapper[4884]: I1202 01:58:37.370122 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.214496 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-b74v6" event={"ID":"2686aabc-d1de-4d56-95db-0b919d055b3d","Type":"ContainerStarted","Data":"24be14cd74684b8efe78f4efdb6cc002f96bdfb7ffe408549db0620403f46436"} Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.215358 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.217735 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"98757586-4157-403c-8a34-f7d28f2451cc","Type":"ContainerStarted","Data":"38c77e578d9a2599b4508ff28f7a7b043df1e1b1369636bfdb22060b9f4f049d"} Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.219061 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"69686a6c-cdef-4fdf-8838-2624f841a85a","Type":"ContainerStarted","Data":"a7434306ef54ec5129024e57360f772e7cac65d9788835fe4193fe0ed268838b"} Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.219088 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"69686a6c-cdef-4fdf-8838-2624f841a85a","Type":"ContainerStarted","Data":"c1a1d3ff248af8626620469e1b48be5a6752079d2d7662c2414d0b2df7e2d98b"} Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.220603 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bf6e71cc-1f65-462c-8234-3c93203d1728","Type":"ContainerStarted","Data":"bf92762a1db3ee254be4c737489b66d5ec6b31d47378a862216842b18ac20a7a"} Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.220680 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="bf6e71cc-1f65-462c-8234-3c93203d1728" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://bf92762a1db3ee254be4c737489b66d5ec6b31d47378a862216842b18ac20a7a" gracePeriod=30 Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.223046 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d4c86db-e57a-4e79-a242-fe623b71f59b","Type":"ContainerStarted","Data":"a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e"} Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.223072 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d4c86db-e57a-4e79-a242-fe623b71f59b","Type":"ContainerStarted","Data":"1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763"} Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.223145 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5d4c86db-e57a-4e79-a242-fe623b71f59b" containerName="nova-metadata-log" containerID="cri-o://1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763" gracePeriod=30 Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.223215 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5d4c86db-e57a-4e79-a242-fe623b71f59b" containerName="nova-metadata-metadata" containerID="cri-o://a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e" gracePeriod=30 Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.234204 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-b74v6" podStartSLOduration=6.234188118 podStartE2EDuration="6.234188118s" podCreationTimestamp="2025-12-02 01:58:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:58:38.23139994 +0000 UTC m=+1334.907236834" watchObservedRunningTime="2025-12-02 01:58:38.234188118 +0000 UTC m=+1334.910025002" Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.251926 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.899789146 podStartE2EDuration="6.251908265s" podCreationTimestamp="2025-12-02 01:58:32 +0000 UTC" firstStartedPulling="2025-12-02 01:58:33.795978796 +0000 UTC m=+1330.471815680" lastFinishedPulling="2025-12-02 01:58:37.148097915 +0000 UTC m=+1333.823934799" observedRunningTime="2025-12-02 01:58:38.243820056 +0000 UTC m=+1334.919656960" watchObservedRunningTime="2025-12-02 01:58:38.251908265 +0000 UTC m=+1334.927745149" Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.265656 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.526365777 podStartE2EDuration="6.265637703s" podCreationTimestamp="2025-12-02 01:58:32 +0000 UTC" firstStartedPulling="2025-12-02 01:58:33.400358701 +0000 UTC m=+1330.076195585" lastFinishedPulling="2025-12-02 01:58:37.139630627 +0000 UTC m=+1333.815467511" observedRunningTime="2025-12-02 01:58:38.256853917 +0000 UTC m=+1334.932690811" watchObservedRunningTime="2025-12-02 01:58:38.265637703 +0000 UTC m=+1334.941474587" Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.277867 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.460753941 podStartE2EDuration="6.277845914s" podCreationTimestamp="2025-12-02 01:58:32 +0000 UTC" firstStartedPulling="2025-12-02 01:58:33.322450022 +0000 UTC m=+1329.998286906" lastFinishedPulling="2025-12-02 01:58:37.139541995 +0000 UTC m=+1333.815378879" observedRunningTime="2025-12-02 01:58:38.269900828 +0000 UTC m=+1334.945737722" watchObservedRunningTime="2025-12-02 01:58:38.277845914 +0000 UTC m=+1334.953682798" Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.302295 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.774137971 podStartE2EDuration="6.302273176s" podCreationTimestamp="2025-12-02 01:58:32 +0000 UTC" firstStartedPulling="2025-12-02 01:58:33.611790219 +0000 UTC m=+1330.287627103" lastFinishedPulling="2025-12-02 01:58:37.139925424 +0000 UTC m=+1333.815762308" observedRunningTime="2025-12-02 01:58:38.290904605 +0000 UTC m=+1334.966741499" watchObservedRunningTime="2025-12-02 01:58:38.302273176 +0000 UTC m=+1334.978110060" Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.814571 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.824581 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d4c86db-e57a-4e79-a242-fe623b71f59b-logs\") pod \"5d4c86db-e57a-4e79-a242-fe623b71f59b\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.824687 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d4c86db-e57a-4e79-a242-fe623b71f59b-config-data\") pod \"5d4c86db-e57a-4e79-a242-fe623b71f59b\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.824791 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jtv5\" (UniqueName: \"kubernetes.io/projected/5d4c86db-e57a-4e79-a242-fe623b71f59b-kube-api-access-8jtv5\") pod \"5d4c86db-e57a-4e79-a242-fe623b71f59b\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.824887 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d4c86db-e57a-4e79-a242-fe623b71f59b-combined-ca-bundle\") pod \"5d4c86db-e57a-4e79-a242-fe623b71f59b\" (UID: \"5d4c86db-e57a-4e79-a242-fe623b71f59b\") " Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.827757 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d4c86db-e57a-4e79-a242-fe623b71f59b-logs" (OuterVolumeSpecName: "logs") pod "5d4c86db-e57a-4e79-a242-fe623b71f59b" (UID: "5d4c86db-e57a-4e79-a242-fe623b71f59b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.831148 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d4c86db-e57a-4e79-a242-fe623b71f59b-kube-api-access-8jtv5" (OuterVolumeSpecName: "kube-api-access-8jtv5") pod "5d4c86db-e57a-4e79-a242-fe623b71f59b" (UID: "5d4c86db-e57a-4e79-a242-fe623b71f59b"). InnerVolumeSpecName "kube-api-access-8jtv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.887836 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d4c86db-e57a-4e79-a242-fe623b71f59b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d4c86db-e57a-4e79-a242-fe623b71f59b" (UID: "5d4c86db-e57a-4e79-a242-fe623b71f59b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.887882 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d4c86db-e57a-4e79-a242-fe623b71f59b-config-data" (OuterVolumeSpecName: "config-data") pod "5d4c86db-e57a-4e79-a242-fe623b71f59b" (UID: "5d4c86db-e57a-4e79-a242-fe623b71f59b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.927551 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d4c86db-e57a-4e79-a242-fe623b71f59b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.927585 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d4c86db-e57a-4e79-a242-fe623b71f59b-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.927595 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d4c86db-e57a-4e79-a242-fe623b71f59b-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:38 crc kubenswrapper[4884]: I1202 01:58:38.927603 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jtv5\" (UniqueName: \"kubernetes.io/projected/5d4c86db-e57a-4e79-a242-fe623b71f59b-kube-api-access-8jtv5\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.235806 4884 generic.go:334] "Generic (PLEG): container finished" podID="5d4c86db-e57a-4e79-a242-fe623b71f59b" containerID="a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e" exitCode=0 Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.235843 4884 generic.go:334] "Generic (PLEG): container finished" podID="5d4c86db-e57a-4e79-a242-fe623b71f59b" containerID="1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763" exitCode=143 Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.235882 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.235946 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d4c86db-e57a-4e79-a242-fe623b71f59b","Type":"ContainerDied","Data":"a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e"} Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.235982 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d4c86db-e57a-4e79-a242-fe623b71f59b","Type":"ContainerDied","Data":"1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763"} Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.235992 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5d4c86db-e57a-4e79-a242-fe623b71f59b","Type":"ContainerDied","Data":"b78b28cecaf9de4f9d0752bd5ba6050cdd9f157f77b053d45d84ba93fe957c96"} Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.236008 4884 scope.go:117] "RemoveContainer" containerID="a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.271451 4884 scope.go:117] "RemoveContainer" containerID="1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.280452 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.298824 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.302733 4884 scope.go:117] "RemoveContainer" containerID="a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e" Dec 02 01:58:39 crc kubenswrapper[4884]: E1202 01:58:39.304428 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e\": container with ID starting with a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e not found: ID does not exist" containerID="a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.304469 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e"} err="failed to get container status \"a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e\": rpc error: code = NotFound desc = could not find container \"a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e\": container with ID starting with a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e not found: ID does not exist" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.304493 4884 scope.go:117] "RemoveContainer" containerID="1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763" Dec 02 01:58:39 crc kubenswrapper[4884]: E1202 01:58:39.306224 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763\": container with ID starting with 1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763 not found: ID does not exist" containerID="1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.306248 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763"} err="failed to get container status \"1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763\": rpc error: code = NotFound desc = could not find container \"1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763\": container with ID starting with 1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763 not found: ID does not exist" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.306263 4884 scope.go:117] "RemoveContainer" containerID="a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.306629 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e"} err="failed to get container status \"a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e\": rpc error: code = NotFound desc = could not find container \"a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e\": container with ID starting with a7ed13e6a011c1f74cb4b120d6af01e0c78d3fcbec48d0b75b223a6b24f2b32e not found: ID does not exist" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.306675 4884 scope.go:117] "RemoveContainer" containerID="1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.307000 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763"} err="failed to get container status \"1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763\": rpc error: code = NotFound desc = could not find container \"1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763\": container with ID starting with 1a75e70847f24548b17658bfdb896a8c5ad8051897bb1c03fe1e2aea1adf4763 not found: ID does not exist" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.312545 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:58:39 crc kubenswrapper[4884]: E1202 01:58:39.313019 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d4c86db-e57a-4e79-a242-fe623b71f59b" containerName="nova-metadata-log" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.313037 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d4c86db-e57a-4e79-a242-fe623b71f59b" containerName="nova-metadata-log" Dec 02 01:58:39 crc kubenswrapper[4884]: E1202 01:58:39.313050 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d4c86db-e57a-4e79-a242-fe623b71f59b" containerName="nova-metadata-metadata" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.313059 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d4c86db-e57a-4e79-a242-fe623b71f59b" containerName="nova-metadata-metadata" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.313247 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d4c86db-e57a-4e79-a242-fe623b71f59b" containerName="nova-metadata-log" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.313278 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d4c86db-e57a-4e79-a242-fe623b71f59b" containerName="nova-metadata-metadata" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.314510 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.321186 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.321414 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.332141 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.335158 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4jn4\" (UniqueName: \"kubernetes.io/projected/2e4be8cb-cb19-4df1-afa9-888f65c32620-kube-api-access-j4jn4\") pod \"nova-metadata-0\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.335231 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.335281 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e4be8cb-cb19-4df1-afa9-888f65c32620-logs\") pod \"nova-metadata-0\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.335309 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-config-data\") pod \"nova-metadata-0\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.335402 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.436188 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.436282 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4jn4\" (UniqueName: \"kubernetes.io/projected/2e4be8cb-cb19-4df1-afa9-888f65c32620-kube-api-access-j4jn4\") pod \"nova-metadata-0\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.436338 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.436376 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e4be8cb-cb19-4df1-afa9-888f65c32620-logs\") pod \"nova-metadata-0\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.436408 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-config-data\") pod \"nova-metadata-0\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.436823 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e4be8cb-cb19-4df1-afa9-888f65c32620-logs\") pod \"nova-metadata-0\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.443632 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.445389 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.447423 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-config-data\") pod \"nova-metadata-0\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.477071 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4jn4\" (UniqueName: \"kubernetes.io/projected/2e4be8cb-cb19-4df1-afa9-888f65c32620-kube-api-access-j4jn4\") pod \"nova-metadata-0\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " pod="openstack/nova-metadata-0" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.626223 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d4c86db-e57a-4e79-a242-fe623b71f59b" path="/var/lib/kubelet/pods/5d4c86db-e57a-4e79-a242-fe623b71f59b/volumes" Dec 02 01:58:39 crc kubenswrapper[4884]: I1202 01:58:39.632526 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 01:58:40 crc kubenswrapper[4884]: I1202 01:58:40.096478 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:58:40 crc kubenswrapper[4884]: W1202 01:58:40.102999 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e4be8cb_cb19_4df1_afa9_888f65c32620.slice/crio-72d197bf38bfd568b75ca3c2e7af369ef4e59f50d34ba765f3ecd4dea82414b4 WatchSource:0}: Error finding container 72d197bf38bfd568b75ca3c2e7af369ef4e59f50d34ba765f3ecd4dea82414b4: Status 404 returned error can't find the container with id 72d197bf38bfd568b75ca3c2e7af369ef4e59f50d34ba765f3ecd4dea82414b4 Dec 02 01:58:40 crc kubenswrapper[4884]: I1202 01:58:40.247222 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e4be8cb-cb19-4df1-afa9-888f65c32620","Type":"ContainerStarted","Data":"72d197bf38bfd568b75ca3c2e7af369ef4e59f50d34ba765f3ecd4dea82414b4"} Dec 02 01:58:41 crc kubenswrapper[4884]: I1202 01:58:41.263042 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e4be8cb-cb19-4df1-afa9-888f65c32620","Type":"ContainerStarted","Data":"daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7"} Dec 02 01:58:41 crc kubenswrapper[4884]: I1202 01:58:41.263370 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e4be8cb-cb19-4df1-afa9-888f65c32620","Type":"ContainerStarted","Data":"5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d"} Dec 02 01:58:41 crc kubenswrapper[4884]: I1202 01:58:41.287778 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.287731393 podStartE2EDuration="2.287731393s" podCreationTimestamp="2025-12-02 01:58:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:58:41.286614746 +0000 UTC m=+1337.962451630" watchObservedRunningTime="2025-12-02 01:58:41.287731393 +0000 UTC m=+1337.963568317" Dec 02 01:58:41 crc kubenswrapper[4884]: I1202 01:58:41.730191 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 01:58:42 crc kubenswrapper[4884]: I1202 01:58:42.278142 4884 generic.go:334] "Generic (PLEG): container finished" podID="5a0a5720-88c1-4adc-9f54-2db975a9e48a" containerID="a4b47a3ed380d29bda07b1cc310be577ed09eece497e9ca981028164b29fc5a4" exitCode=0 Dec 02 01:58:42 crc kubenswrapper[4884]: I1202 01:58:42.278231 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x695l" event={"ID":"5a0a5720-88c1-4adc-9f54-2db975a9e48a","Type":"ContainerDied","Data":"a4b47a3ed380d29bda07b1cc310be577ed09eece497e9ca981028164b29fc5a4"} Dec 02 01:58:42 crc kubenswrapper[4884]: I1202 01:58:42.284615 4884 generic.go:334] "Generic (PLEG): container finished" podID="e74647f6-5b85-47c2-9f26-27465f963095" containerID="e54c5930d9ab03a015536b995f4926f387f59c08dd7848b2af7b37b877f67176" exitCode=0 Dec 02 01:58:42 crc kubenswrapper[4884]: I1202 01:58:42.284928 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="1583598a-ff88-4fea-bc25-5ff097443e7f" containerName="kube-state-metrics" containerID="cri-o://f0935a91358e557e628541d3800d9ca9311441f0dd6c3fad9a949bfe6b0d8921" gracePeriod=30 Dec 02 01:58:42 crc kubenswrapper[4884]: I1202 01:58:42.285275 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-kcpzh" event={"ID":"e74647f6-5b85-47c2-9f26-27465f963095","Type":"ContainerDied","Data":"e54c5930d9ab03a015536b995f4926f387f59c08dd7848b2af7b37b877f67176"} Dec 02 01:58:42 crc kubenswrapper[4884]: I1202 01:58:42.666709 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 01:58:42 crc kubenswrapper[4884]: I1202 01:58:42.667371 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 01:58:42 crc kubenswrapper[4884]: I1202 01:58:42.721723 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 02 01:58:42 crc kubenswrapper[4884]: I1202 01:58:42.722041 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 02 01:58:42 crc kubenswrapper[4884]: I1202 01:58:42.752407 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 02 01:58:42 crc kubenswrapper[4884]: I1202 01:58:42.916293 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.002253 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.009390 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mk2jc\" (UniqueName: \"kubernetes.io/projected/1583598a-ff88-4fea-bc25-5ff097443e7f-kube-api-access-mk2jc\") pod \"1583598a-ff88-4fea-bc25-5ff097443e7f\" (UID: \"1583598a-ff88-4fea-bc25-5ff097443e7f\") " Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.023973 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1583598a-ff88-4fea-bc25-5ff097443e7f-kube-api-access-mk2jc" (OuterVolumeSpecName: "kube-api-access-mk2jc") pod "1583598a-ff88-4fea-bc25-5ff097443e7f" (UID: "1583598a-ff88-4fea-bc25-5ff097443e7f"). InnerVolumeSpecName "kube-api-access-mk2jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.069937 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.110682 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mk2jc\" (UniqueName: \"kubernetes.io/projected/1583598a-ff88-4fea-bc25-5ff097443e7f-kube-api-access-mk2jc\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.143067 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-cbbd4"] Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.143282 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" podUID="91335ed2-246d-42b7-9d69-7ac7aada058b" containerName="dnsmasq-dns" containerID="cri-o://d27a0f434086c0b10da27c069a025623ffb6e9ac1fe49ff8d19f8ade64719295" gracePeriod=10 Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.302124 4884 generic.go:334] "Generic (PLEG): container finished" podID="1583598a-ff88-4fea-bc25-5ff097443e7f" containerID="f0935a91358e557e628541d3800d9ca9311441f0dd6c3fad9a949bfe6b0d8921" exitCode=2 Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.302267 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.303172 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1583598a-ff88-4fea-bc25-5ff097443e7f","Type":"ContainerDied","Data":"f0935a91358e557e628541d3800d9ca9311441f0dd6c3fad9a949bfe6b0d8921"} Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.303205 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1583598a-ff88-4fea-bc25-5ff097443e7f","Type":"ContainerDied","Data":"7259233685c2d6fd6d95e04abfe630073fe780410305d7d91fcc4bd751a91e67"} Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.303225 4884 scope.go:117] "RemoveContainer" containerID="f0935a91358e557e628541d3800d9ca9311441f0dd6c3fad9a949bfe6b0d8921" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.315951 4884 generic.go:334] "Generic (PLEG): container finished" podID="91335ed2-246d-42b7-9d69-7ac7aada058b" containerID="d27a0f434086c0b10da27c069a025623ffb6e9ac1fe49ff8d19f8ade64719295" exitCode=0 Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.317344 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" event={"ID":"91335ed2-246d-42b7-9d69-7ac7aada058b","Type":"ContainerDied","Data":"d27a0f434086c0b10da27c069a025623ffb6e9ac1fe49ff8d19f8ade64719295"} Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.350829 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.368470 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.375792 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.385255 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 01:58:43 crc kubenswrapper[4884]: E1202 01:58:43.385699 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1583598a-ff88-4fea-bc25-5ff097443e7f" containerName="kube-state-metrics" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.385714 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="1583598a-ff88-4fea-bc25-5ff097443e7f" containerName="kube-state-metrics" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.385919 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="1583598a-ff88-4fea-bc25-5ff097443e7f" containerName="kube-state-metrics" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.386580 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.391432 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.391761 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.400885 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.514967 4884 scope.go:117] "RemoveContainer" containerID="f0935a91358e557e628541d3800d9ca9311441f0dd6c3fad9a949bfe6b0d8921" Dec 02 01:58:43 crc kubenswrapper[4884]: E1202 01:58:43.516628 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0935a91358e557e628541d3800d9ca9311441f0dd6c3fad9a949bfe6b0d8921\": container with ID starting with f0935a91358e557e628541d3800d9ca9311441f0dd6c3fad9a949bfe6b0d8921 not found: ID does not exist" containerID="f0935a91358e557e628541d3800d9ca9311441f0dd6c3fad9a949bfe6b0d8921" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.516694 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0935a91358e557e628541d3800d9ca9311441f0dd6c3fad9a949bfe6b0d8921"} err="failed to get container status \"f0935a91358e557e628541d3800d9ca9311441f0dd6c3fad9a949bfe6b0d8921\": rpc error: code = NotFound desc = could not find container \"f0935a91358e557e628541d3800d9ca9311441f0dd6c3fad9a949bfe6b0d8921\": container with ID starting with f0935a91358e557e628541d3800d9ca9311441f0dd6c3fad9a949bfe6b0d8921 not found: ID does not exist" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.521525 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnjwn\" (UniqueName: \"kubernetes.io/projected/9537581d-d462-4e9d-992e-15c9754c3fb3-kube-api-access-dnjwn\") pod \"kube-state-metrics-0\" (UID: \"9537581d-d462-4e9d-992e-15c9754c3fb3\") " pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.521566 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9537581d-d462-4e9d-992e-15c9754c3fb3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9537581d-d462-4e9d-992e-15c9754c3fb3\") " pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.521655 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9537581d-d462-4e9d-992e-15c9754c3fb3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9537581d-d462-4e9d-992e-15c9754c3fb3\") " pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.521878 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9537581d-d462-4e9d-992e-15c9754c3fb3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9537581d-d462-4e9d-992e-15c9754c3fb3\") " pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.623823 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnjwn\" (UniqueName: \"kubernetes.io/projected/9537581d-d462-4e9d-992e-15c9754c3fb3-kube-api-access-dnjwn\") pod \"kube-state-metrics-0\" (UID: \"9537581d-d462-4e9d-992e-15c9754c3fb3\") " pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.623860 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9537581d-d462-4e9d-992e-15c9754c3fb3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9537581d-d462-4e9d-992e-15c9754c3fb3\") " pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.623895 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9537581d-d462-4e9d-992e-15c9754c3fb3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9537581d-d462-4e9d-992e-15c9754c3fb3\") " pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.624592 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9537581d-d462-4e9d-992e-15c9754c3fb3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9537581d-d462-4e9d-992e-15c9754c3fb3\") " pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.638350 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1583598a-ff88-4fea-bc25-5ff097443e7f" path="/var/lib/kubelet/pods/1583598a-ff88-4fea-bc25-5ff097443e7f/volumes" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.639151 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9537581d-d462-4e9d-992e-15c9754c3fb3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9537581d-d462-4e9d-992e-15c9754c3fb3\") " pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.642512 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9537581d-d462-4e9d-992e-15c9754c3fb3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9537581d-d462-4e9d-992e-15c9754c3fb3\") " pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.650610 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9537581d-d462-4e9d-992e-15c9754c3fb3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9537581d-d462-4e9d-992e-15c9754c3fb3\") " pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.653883 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnjwn\" (UniqueName: \"kubernetes.io/projected/9537581d-d462-4e9d-992e-15c9754c3fb3-kube-api-access-dnjwn\") pod \"kube-state-metrics-0\" (UID: \"9537581d-d462-4e9d-992e-15c9754c3fb3\") " pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.716362 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.750141 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="69686a6c-cdef-4fdf-8838-2624f841a85a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.750165 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="69686a6c-cdef-4fdf-8838-2624f841a85a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.201:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.846192 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.930598 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-config\") pod \"91335ed2-246d-42b7-9d69-7ac7aada058b\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.930859 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-dns-swift-storage-0\") pod \"91335ed2-246d-42b7-9d69-7ac7aada058b\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.930939 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-ovsdbserver-sb\") pod \"91335ed2-246d-42b7-9d69-7ac7aada058b\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.930957 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qb4vp\" (UniqueName: \"kubernetes.io/projected/91335ed2-246d-42b7-9d69-7ac7aada058b-kube-api-access-qb4vp\") pod \"91335ed2-246d-42b7-9d69-7ac7aada058b\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.931058 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-dns-svc\") pod \"91335ed2-246d-42b7-9d69-7ac7aada058b\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.931104 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-ovsdbserver-nb\") pod \"91335ed2-246d-42b7-9d69-7ac7aada058b\" (UID: \"91335ed2-246d-42b7-9d69-7ac7aada058b\") " Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.938473 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91335ed2-246d-42b7-9d69-7ac7aada058b-kube-api-access-qb4vp" (OuterVolumeSpecName: "kube-api-access-qb4vp") pod "91335ed2-246d-42b7-9d69-7ac7aada058b" (UID: "91335ed2-246d-42b7-9d69-7ac7aada058b"). InnerVolumeSpecName "kube-api-access-qb4vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:58:43 crc kubenswrapper[4884]: I1202 01:58:43.952995 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.002166 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.007160 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "91335ed2-246d-42b7-9d69-7ac7aada058b" (UID: "91335ed2-246d-42b7-9d69-7ac7aada058b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.009766 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "91335ed2-246d-42b7-9d69-7ac7aada058b" (UID: "91335ed2-246d-42b7-9d69-7ac7aada058b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.010429 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-config" (OuterVolumeSpecName: "config") pod "91335ed2-246d-42b7-9d69-7ac7aada058b" (UID: "91335ed2-246d-42b7-9d69-7ac7aada058b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.035199 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.035240 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qb4vp\" (UniqueName: \"kubernetes.io/projected/91335ed2-246d-42b7-9d69-7ac7aada058b-kube-api-access-qb4vp\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.035257 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.035268 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.067064 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "91335ed2-246d-42b7-9d69-7ac7aada058b" (UID: "91335ed2-246d-42b7-9d69-7ac7aada058b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.079209 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "91335ed2-246d-42b7-9d69-7ac7aada058b" (UID: "91335ed2-246d-42b7-9d69-7ac7aada058b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.136883 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-scripts\") pod \"e74647f6-5b85-47c2-9f26-27465f963095\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.136921 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-combined-ca-bundle\") pod \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.136950 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-config-data\") pod \"e74647f6-5b85-47c2-9f26-27465f963095\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.137022 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsblp\" (UniqueName: \"kubernetes.io/projected/5a0a5720-88c1-4adc-9f54-2db975a9e48a-kube-api-access-wsblp\") pod \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.137053 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-config-data\") pod \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.137067 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-scripts\") pod \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\" (UID: \"5a0a5720-88c1-4adc-9f54-2db975a9e48a\") " Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.137103 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdbkt\" (UniqueName: \"kubernetes.io/projected/e74647f6-5b85-47c2-9f26-27465f963095-kube-api-access-kdbkt\") pod \"e74647f6-5b85-47c2-9f26-27465f963095\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.137166 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-combined-ca-bundle\") pod \"e74647f6-5b85-47c2-9f26-27465f963095\" (UID: \"e74647f6-5b85-47c2-9f26-27465f963095\") " Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.137657 4884 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.137669 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/91335ed2-246d-42b7-9d69-7ac7aada058b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.154006 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-scripts" (OuterVolumeSpecName: "scripts") pod "e74647f6-5b85-47c2-9f26-27465f963095" (UID: "e74647f6-5b85-47c2-9f26-27465f963095"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.159026 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e74647f6-5b85-47c2-9f26-27465f963095-kube-api-access-kdbkt" (OuterVolumeSpecName: "kube-api-access-kdbkt") pod "e74647f6-5b85-47c2-9f26-27465f963095" (UID: "e74647f6-5b85-47c2-9f26-27465f963095"). InnerVolumeSpecName "kube-api-access-kdbkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.159083 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a0a5720-88c1-4adc-9f54-2db975a9e48a-kube-api-access-wsblp" (OuterVolumeSpecName: "kube-api-access-wsblp") pod "5a0a5720-88c1-4adc-9f54-2db975a9e48a" (UID: "5a0a5720-88c1-4adc-9f54-2db975a9e48a"). InnerVolumeSpecName "kube-api-access-wsblp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.163558 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-scripts" (OuterVolumeSpecName: "scripts") pod "5a0a5720-88c1-4adc-9f54-2db975a9e48a" (UID: "5a0a5720-88c1-4adc-9f54-2db975a9e48a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.204892 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-config-data" (OuterVolumeSpecName: "config-data") pod "5a0a5720-88c1-4adc-9f54-2db975a9e48a" (UID: "5a0a5720-88c1-4adc-9f54-2db975a9e48a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.215423 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e74647f6-5b85-47c2-9f26-27465f963095" (UID: "e74647f6-5b85-47c2-9f26-27465f963095"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.223932 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-config-data" (OuterVolumeSpecName: "config-data") pod "e74647f6-5b85-47c2-9f26-27465f963095" (UID: "e74647f6-5b85-47c2-9f26-27465f963095"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.224362 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a0a5720-88c1-4adc-9f54-2db975a9e48a" (UID: "5a0a5720-88c1-4adc-9f54-2db975a9e48a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.239534 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.240025 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.240042 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdbkt\" (UniqueName: \"kubernetes.io/projected/e74647f6-5b85-47c2-9f26-27465f963095-kube-api-access-kdbkt\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.240052 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.240062 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.240070 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a0a5720-88c1-4adc-9f54-2db975a9e48a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.240078 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e74647f6-5b85-47c2-9f26-27465f963095-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.240086 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsblp\" (UniqueName: \"kubernetes.io/projected/5a0a5720-88c1-4adc-9f54-2db975a9e48a-kube-api-access-wsblp\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.281938 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.282830 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="ceilometer-central-agent" containerID="cri-o://2909efeea26d7f14bc4d65944e00b64dceb7904c66064476b993decf02b781db" gracePeriod=30 Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.283857 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="sg-core" containerID="cri-o://a026cb2e4e0d8ebc6078faa721738be77d0dabd0ba1f95c630b264b1b84a471f" gracePeriod=30 Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.283872 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="proxy-httpd" containerID="cri-o://e822efae88553c322dcf0063bcb0396d1dd3f2a9969a1751f69a70177c6a82e6" gracePeriod=30 Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.283881 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="ceilometer-notification-agent" containerID="cri-o://c9e19a9dfbb3d9d2019696ffd26fdfa8e5652b4d1ba8ce862f6be658dae9baf9" gracePeriod=30 Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.327629 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x695l" event={"ID":"5a0a5720-88c1-4adc-9f54-2db975a9e48a","Type":"ContainerDied","Data":"c09c82e60c9ede7ff70afa2ba6993db2eeab484eca31542a4bfd8bffd93c974d"} Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.327896 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c09c82e60c9ede7ff70afa2ba6993db2eeab484eca31542a4bfd8bffd93c974d" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.327943 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x695l" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.334186 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-kcpzh" event={"ID":"e74647f6-5b85-47c2-9f26-27465f963095","Type":"ContainerDied","Data":"108530649f545fcc5e409e072f9dbafbd2a34449bfbe6ee518039bbe6cb7ac52"} Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.334356 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="108530649f545fcc5e409e072f9dbafbd2a34449bfbe6ee518039bbe6cb7ac52" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.334440 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-kcpzh" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.346975 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.351111 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-cbbd4" event={"ID":"91335ed2-246d-42b7-9d69-7ac7aada058b","Type":"ContainerDied","Data":"bb80f41bf6b5d2c09f29fdde80c97df4149f45281103fb9609281a746c663022"} Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.351316 4884 scope.go:117] "RemoveContainer" containerID="d27a0f434086c0b10da27c069a025623ffb6e9ac1fe49ff8d19f8ade64719295" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.398928 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.412647 4884 scope.go:117] "RemoveContainer" containerID="913e9170ef6de8e043f6272ead9c298935c9ab5c5862bd5ca39a2505fd748383" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.448805 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 02 01:58:44 crc kubenswrapper[4884]: E1202 01:58:44.449491 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e74647f6-5b85-47c2-9f26-27465f963095" containerName="nova-cell1-conductor-db-sync" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.449512 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="e74647f6-5b85-47c2-9f26-27465f963095" containerName="nova-cell1-conductor-db-sync" Dec 02 01:58:44 crc kubenswrapper[4884]: E1202 01:58:44.449540 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91335ed2-246d-42b7-9d69-7ac7aada058b" containerName="init" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.449564 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="91335ed2-246d-42b7-9d69-7ac7aada058b" containerName="init" Dec 02 01:58:44 crc kubenswrapper[4884]: E1202 01:58:44.449579 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91335ed2-246d-42b7-9d69-7ac7aada058b" containerName="dnsmasq-dns" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.449587 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="91335ed2-246d-42b7-9d69-7ac7aada058b" containerName="dnsmasq-dns" Dec 02 01:58:44 crc kubenswrapper[4884]: E1202 01:58:44.449605 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a0a5720-88c1-4adc-9f54-2db975a9e48a" containerName="nova-manage" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.449612 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a0a5720-88c1-4adc-9f54-2db975a9e48a" containerName="nova-manage" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.449930 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a0a5720-88c1-4adc-9f54-2db975a9e48a" containerName="nova-manage" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.449970 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="e74647f6-5b85-47c2-9f26-27465f963095" containerName="nova-cell1-conductor-db-sync" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.450010 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="91335ed2-246d-42b7-9d69-7ac7aada058b" containerName="dnsmasq-dns" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.451011 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.454332 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.464839 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.477297 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-cbbd4"] Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.493099 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-cbbd4"] Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.570095 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpbv7\" (UniqueName: \"kubernetes.io/projected/5993c2f9-ca6d-454f-ae88-08b91e419b3f-kube-api-access-wpbv7\") pod \"nova-cell1-conductor-0\" (UID: \"5993c2f9-ca6d-454f-ae88-08b91e419b3f\") " pod="openstack/nova-cell1-conductor-0" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.570138 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5993c2f9-ca6d-454f-ae88-08b91e419b3f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5993c2f9-ca6d-454f-ae88-08b91e419b3f\") " pod="openstack/nova-cell1-conductor-0" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.570172 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5993c2f9-ca6d-454f-ae88-08b91e419b3f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5993c2f9-ca6d-454f-ae88-08b91e419b3f\") " pod="openstack/nova-cell1-conductor-0" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.616141 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.630237 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.630520 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="69686a6c-cdef-4fdf-8838-2624f841a85a" containerName="nova-api-log" containerID="cri-o://c1a1d3ff248af8626620469e1b48be5a6752079d2d7662c2414d0b2df7e2d98b" gracePeriod=30 Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.630586 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="69686a6c-cdef-4fdf-8838-2624f841a85a" containerName="nova-api-api" containerID="cri-o://a7434306ef54ec5129024e57360f772e7cac65d9788835fe4193fe0ed268838b" gracePeriod=30 Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.633168 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.633576 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.658121 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.672923 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5993c2f9-ca6d-454f-ae88-08b91e419b3f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5993c2f9-ca6d-454f-ae88-08b91e419b3f\") " pod="openstack/nova-cell1-conductor-0" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.672958 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpbv7\" (UniqueName: \"kubernetes.io/projected/5993c2f9-ca6d-454f-ae88-08b91e419b3f-kube-api-access-wpbv7\") pod \"nova-cell1-conductor-0\" (UID: \"5993c2f9-ca6d-454f-ae88-08b91e419b3f\") " pod="openstack/nova-cell1-conductor-0" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.672992 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5993c2f9-ca6d-454f-ae88-08b91e419b3f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5993c2f9-ca6d-454f-ae88-08b91e419b3f\") " pod="openstack/nova-cell1-conductor-0" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.680156 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5993c2f9-ca6d-454f-ae88-08b91e419b3f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5993c2f9-ca6d-454f-ae88-08b91e419b3f\") " pod="openstack/nova-cell1-conductor-0" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.682136 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5993c2f9-ca6d-454f-ae88-08b91e419b3f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5993c2f9-ca6d-454f-ae88-08b91e419b3f\") " pod="openstack/nova-cell1-conductor-0" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.698140 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpbv7\" (UniqueName: \"kubernetes.io/projected/5993c2f9-ca6d-454f-ae88-08b91e419b3f-kube-api-access-wpbv7\") pod \"nova-cell1-conductor-0\" (UID: \"5993c2f9-ca6d-454f-ae88-08b91e419b3f\") " pod="openstack/nova-cell1-conductor-0" Dec 02 01:58:44 crc kubenswrapper[4884]: I1202 01:58:44.780328 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 02 01:58:45 crc kubenswrapper[4884]: I1202 01:58:45.335049 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 02 01:58:45 crc kubenswrapper[4884]: I1202 01:58:45.351224 4884 generic.go:334] "Generic (PLEG): container finished" podID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerID="e822efae88553c322dcf0063bcb0396d1dd3f2a9969a1751f69a70177c6a82e6" exitCode=0 Dec 02 01:58:45 crc kubenswrapper[4884]: I1202 01:58:45.351250 4884 generic.go:334] "Generic (PLEG): container finished" podID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerID="a026cb2e4e0d8ebc6078faa721738be77d0dabd0ba1f95c630b264b1b84a471f" exitCode=2 Dec 02 01:58:45 crc kubenswrapper[4884]: I1202 01:58:45.351258 4884 generic.go:334] "Generic (PLEG): container finished" podID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerID="2909efeea26d7f14bc4d65944e00b64dceb7904c66064476b993decf02b781db" exitCode=0 Dec 02 01:58:45 crc kubenswrapper[4884]: I1202 01:58:45.351271 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6e289da-5ac9-4191-8bf1-2611786733d5","Type":"ContainerDied","Data":"e822efae88553c322dcf0063bcb0396d1dd3f2a9969a1751f69a70177c6a82e6"} Dec 02 01:58:45 crc kubenswrapper[4884]: I1202 01:58:45.351313 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6e289da-5ac9-4191-8bf1-2611786733d5","Type":"ContainerDied","Data":"a026cb2e4e0d8ebc6078faa721738be77d0dabd0ba1f95c630b264b1b84a471f"} Dec 02 01:58:45 crc kubenswrapper[4884]: I1202 01:58:45.351326 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6e289da-5ac9-4191-8bf1-2611786733d5","Type":"ContainerDied","Data":"2909efeea26d7f14bc4d65944e00b64dceb7904c66064476b993decf02b781db"} Dec 02 01:58:45 crc kubenswrapper[4884]: I1202 01:58:45.352701 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9537581d-d462-4e9d-992e-15c9754c3fb3","Type":"ContainerStarted","Data":"56de36d44751fbccb89938de764c43a413a7a4753ccaf454e6e83897f92f37c8"} Dec 02 01:58:45 crc kubenswrapper[4884]: I1202 01:58:45.352730 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9537581d-d462-4e9d-992e-15c9754c3fb3","Type":"ContainerStarted","Data":"ba605f7e7e9b5b9b6af5ebcb7ecf2a11972c027f762dcfa5bb5ae7569ac8c837"} Dec 02 01:58:45 crc kubenswrapper[4884]: I1202 01:58:45.355463 4884 generic.go:334] "Generic (PLEG): container finished" podID="69686a6c-cdef-4fdf-8838-2624f841a85a" containerID="c1a1d3ff248af8626620469e1b48be5a6752079d2d7662c2414d0b2df7e2d98b" exitCode=143 Dec 02 01:58:45 crc kubenswrapper[4884]: I1202 01:58:45.355506 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"69686a6c-cdef-4fdf-8838-2624f841a85a","Type":"ContainerDied","Data":"c1a1d3ff248af8626620469e1b48be5a6752079d2d7662c2414d0b2df7e2d98b"} Dec 02 01:58:45 crc kubenswrapper[4884]: I1202 01:58:45.356778 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5993c2f9-ca6d-454f-ae88-08b91e419b3f","Type":"ContainerStarted","Data":"34fb416265fd9df2ad15df016bf70121911d514520cf0550886ccc211deca26a"} Dec 02 01:58:45 crc kubenswrapper[4884]: I1202 01:58:45.373199 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.01448205 podStartE2EDuration="2.373185526s" podCreationTimestamp="2025-12-02 01:58:43 +0000 UTC" firstStartedPulling="2025-12-02 01:58:44.365802842 +0000 UTC m=+1341.041639726" lastFinishedPulling="2025-12-02 01:58:44.724506318 +0000 UTC m=+1341.400343202" observedRunningTime="2025-12-02 01:58:45.369638409 +0000 UTC m=+1342.045475293" watchObservedRunningTime="2025-12-02 01:58:45.373185526 +0000 UTC m=+1342.049022410" Dec 02 01:58:45 crc kubenswrapper[4884]: I1202 01:58:45.628985 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91335ed2-246d-42b7-9d69-7ac7aada058b" path="/var/lib/kubelet/pods/91335ed2-246d-42b7-9d69-7ac7aada058b/volumes" Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.383392 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5993c2f9-ca6d-454f-ae88-08b91e419b3f","Type":"ContainerStarted","Data":"1836beaa5836ce43bfc8a63a170ef17acfec03bd8ace63e8994d55a1712a0233"} Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.383791 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.386733 4884 generic.go:334] "Generic (PLEG): container finished" podID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerID="c9e19a9dfbb3d9d2019696ffd26fdfa8e5652b4d1ba8ce862f6be658dae9baf9" exitCode=0 Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.386901 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="98757586-4157-403c-8a34-f7d28f2451cc" containerName="nova-scheduler-scheduler" containerID="cri-o://38c77e578d9a2599b4508ff28f7a7b043df1e1b1369636bfdb22060b9f4f049d" gracePeriod=30 Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.387104 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6e289da-5ac9-4191-8bf1-2611786733d5","Type":"ContainerDied","Data":"c9e19a9dfbb3d9d2019696ffd26fdfa8e5652b4d1ba8ce862f6be658dae9baf9"} Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.387222 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2e4be8cb-cb19-4df1-afa9-888f65c32620" containerName="nova-metadata-log" containerID="cri-o://5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d" gracePeriod=30 Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.387298 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2e4be8cb-cb19-4df1-afa9-888f65c32620" containerName="nova-metadata-metadata" containerID="cri-o://daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7" gracePeriod=30 Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.387441 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.408175 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.408157469 podStartE2EDuration="2.408157469s" podCreationTimestamp="2025-12-02 01:58:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:58:46.399038895 +0000 UTC m=+1343.074875779" watchObservedRunningTime="2025-12-02 01:58:46.408157469 +0000 UTC m=+1343.083994353" Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.775506 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.950265 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6e289da-5ac9-4191-8bf1-2611786733d5-run-httpd\") pod \"d6e289da-5ac9-4191-8bf1-2611786733d5\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.950622 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-combined-ca-bundle\") pod \"d6e289da-5ac9-4191-8bf1-2611786733d5\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.950717 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-scripts\") pod \"d6e289da-5ac9-4191-8bf1-2611786733d5\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.950755 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-sg-core-conf-yaml\") pod \"d6e289da-5ac9-4191-8bf1-2611786733d5\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.950772 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6e289da-5ac9-4191-8bf1-2611786733d5-log-httpd\") pod \"d6e289da-5ac9-4191-8bf1-2611786733d5\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.950881 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qj7kg\" (UniqueName: \"kubernetes.io/projected/d6e289da-5ac9-4191-8bf1-2611786733d5-kube-api-access-qj7kg\") pod \"d6e289da-5ac9-4191-8bf1-2611786733d5\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.950928 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-config-data\") pod \"d6e289da-5ac9-4191-8bf1-2611786733d5\" (UID: \"d6e289da-5ac9-4191-8bf1-2611786733d5\") " Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.953094 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6e289da-5ac9-4191-8bf1-2611786733d5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d6e289da-5ac9-4191-8bf1-2611786733d5" (UID: "d6e289da-5ac9-4191-8bf1-2611786733d5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.955042 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6e289da-5ac9-4191-8bf1-2611786733d5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d6e289da-5ac9-4191-8bf1-2611786733d5" (UID: "d6e289da-5ac9-4191-8bf1-2611786733d5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.961085 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-scripts" (OuterVolumeSpecName: "scripts") pod "d6e289da-5ac9-4191-8bf1-2611786733d5" (UID: "d6e289da-5ac9-4191-8bf1-2611786733d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.966160 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6e289da-5ac9-4191-8bf1-2611786733d5-kube-api-access-qj7kg" (OuterVolumeSpecName: "kube-api-access-qj7kg") pod "d6e289da-5ac9-4191-8bf1-2611786733d5" (UID: "d6e289da-5ac9-4191-8bf1-2611786733d5"). InnerVolumeSpecName "kube-api-access-qj7kg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:58:46 crc kubenswrapper[4884]: I1202 01:58:46.987828 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d6e289da-5ac9-4191-8bf1-2611786733d5" (UID: "d6e289da-5ac9-4191-8bf1-2611786733d5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.053610 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.053673 4884 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.053688 4884 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6e289da-5ac9-4191-8bf1-2611786733d5-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.053701 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qj7kg\" (UniqueName: \"kubernetes.io/projected/d6e289da-5ac9-4191-8bf1-2611786733d5-kube-api-access-qj7kg\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.053762 4884 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6e289da-5ac9-4191-8bf1-2611786733d5-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.068620 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-config-data" (OuterVolumeSpecName: "config-data") pod "d6e289da-5ac9-4191-8bf1-2611786733d5" (UID: "d6e289da-5ac9-4191-8bf1-2611786733d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.073268 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.078995 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6e289da-5ac9-4191-8bf1-2611786733d5" (UID: "d6e289da-5ac9-4191-8bf1-2611786733d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.155517 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.155602 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e289da-5ac9-4191-8bf1-2611786733d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.256385 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-combined-ca-bundle\") pod \"2e4be8cb-cb19-4df1-afa9-888f65c32620\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.256696 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4jn4\" (UniqueName: \"kubernetes.io/projected/2e4be8cb-cb19-4df1-afa9-888f65c32620-kube-api-access-j4jn4\") pod \"2e4be8cb-cb19-4df1-afa9-888f65c32620\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.257129 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e4be8cb-cb19-4df1-afa9-888f65c32620-logs\") pod \"2e4be8cb-cb19-4df1-afa9-888f65c32620\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.257385 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e4be8cb-cb19-4df1-afa9-888f65c32620-logs" (OuterVolumeSpecName: "logs") pod "2e4be8cb-cb19-4df1-afa9-888f65c32620" (UID: "2e4be8cb-cb19-4df1-afa9-888f65c32620"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.257422 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-nova-metadata-tls-certs\") pod \"2e4be8cb-cb19-4df1-afa9-888f65c32620\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.257456 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-config-data\") pod \"2e4be8cb-cb19-4df1-afa9-888f65c32620\" (UID: \"2e4be8cb-cb19-4df1-afa9-888f65c32620\") " Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.257839 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2e4be8cb-cb19-4df1-afa9-888f65c32620-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.260073 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e4be8cb-cb19-4df1-afa9-888f65c32620-kube-api-access-j4jn4" (OuterVolumeSpecName: "kube-api-access-j4jn4") pod "2e4be8cb-cb19-4df1-afa9-888f65c32620" (UID: "2e4be8cb-cb19-4df1-afa9-888f65c32620"). InnerVolumeSpecName "kube-api-access-j4jn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.283285 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e4be8cb-cb19-4df1-afa9-888f65c32620" (UID: "2e4be8cb-cb19-4df1-afa9-888f65c32620"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.291945 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-config-data" (OuterVolumeSpecName: "config-data") pod "2e4be8cb-cb19-4df1-afa9-888f65c32620" (UID: "2e4be8cb-cb19-4df1-afa9-888f65c32620"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.323539 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "2e4be8cb-cb19-4df1-afa9-888f65c32620" (UID: "2e4be8cb-cb19-4df1-afa9-888f65c32620"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.360335 4884 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.360390 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.360410 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e4be8cb-cb19-4df1-afa9-888f65c32620-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.360427 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4jn4\" (UniqueName: \"kubernetes.io/projected/2e4be8cb-cb19-4df1-afa9-888f65c32620-kube-api-access-j4jn4\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.409029 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6e289da-5ac9-4191-8bf1-2611786733d5","Type":"ContainerDied","Data":"7d7dd89f0fde1dad0ed47c5c6271f9adc44dc420a7f156c81942e28406bbeee2"} Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.409080 4884 scope.go:117] "RemoveContainer" containerID="e822efae88553c322dcf0063bcb0396d1dd3f2a9969a1751f69a70177c6a82e6" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.409200 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.414039 4884 generic.go:334] "Generic (PLEG): container finished" podID="2e4be8cb-cb19-4df1-afa9-888f65c32620" containerID="daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7" exitCode=0 Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.414074 4884 generic.go:334] "Generic (PLEG): container finished" podID="2e4be8cb-cb19-4df1-afa9-888f65c32620" containerID="5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d" exitCode=143 Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.415453 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.416545 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e4be8cb-cb19-4df1-afa9-888f65c32620","Type":"ContainerDied","Data":"daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7"} Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.416600 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e4be8cb-cb19-4df1-afa9-888f65c32620","Type":"ContainerDied","Data":"5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d"} Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.416619 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2e4be8cb-cb19-4df1-afa9-888f65c32620","Type":"ContainerDied","Data":"72d197bf38bfd568b75ca3c2e7af369ef4e59f50d34ba765f3ecd4dea82414b4"} Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.438829 4884 scope.go:117] "RemoveContainer" containerID="a026cb2e4e0d8ebc6078faa721738be77d0dabd0ba1f95c630b264b1b84a471f" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.449777 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.456425 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.480103 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.494365 4884 scope.go:117] "RemoveContainer" containerID="c9e19a9dfbb3d9d2019696ffd26fdfa8e5652b4d1ba8ce862f6be658dae9baf9" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.494729 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.522426 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:58:47 crc kubenswrapper[4884]: E1202 01:58:47.522876 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="sg-core" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.522893 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="sg-core" Dec 02 01:58:47 crc kubenswrapper[4884]: E1202 01:58:47.522926 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="ceilometer-central-agent" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.522933 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="ceilometer-central-agent" Dec 02 01:58:47 crc kubenswrapper[4884]: E1202 01:58:47.522954 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e4be8cb-cb19-4df1-afa9-888f65c32620" containerName="nova-metadata-metadata" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.522960 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e4be8cb-cb19-4df1-afa9-888f65c32620" containerName="nova-metadata-metadata" Dec 02 01:58:47 crc kubenswrapper[4884]: E1202 01:58:47.522973 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="ceilometer-notification-agent" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.522978 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="ceilometer-notification-agent" Dec 02 01:58:47 crc kubenswrapper[4884]: E1202 01:58:47.522989 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e4be8cb-cb19-4df1-afa9-888f65c32620" containerName="nova-metadata-log" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.522994 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e4be8cb-cb19-4df1-afa9-888f65c32620" containerName="nova-metadata-log" Dec 02 01:58:47 crc kubenswrapper[4884]: E1202 01:58:47.523006 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="proxy-httpd" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.523030 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="proxy-httpd" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.523232 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e4be8cb-cb19-4df1-afa9-888f65c32620" containerName="nova-metadata-log" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.523245 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="ceilometer-notification-agent" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.523258 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e4be8cb-cb19-4df1-afa9-888f65c32620" containerName="nova-metadata-metadata" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.523275 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="ceilometer-central-agent" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.523282 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="proxy-httpd" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.523291 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" containerName="sg-core" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.525221 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.534674 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.536283 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.551203 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.551252 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.555898 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.556031 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.556203 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.556239 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.556318 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.561148 4884 scope.go:117] "RemoveContainer" containerID="2909efeea26d7f14bc4d65944e00b64dceb7904c66064476b993decf02b781db" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.603322 4884 scope.go:117] "RemoveContainer" containerID="daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.623811 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e4be8cb-cb19-4df1-afa9-888f65c32620" path="/var/lib/kubelet/pods/2e4be8cb-cb19-4df1-afa9-888f65c32620/volumes" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.624650 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6e289da-5ac9-4191-8bf1-2611786733d5" path="/var/lib/kubelet/pods/d6e289da-5ac9-4191-8bf1-2611786733d5/volumes" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.628309 4884 scope.go:117] "RemoveContainer" containerID="5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.647251 4884 scope.go:117] "RemoveContainer" containerID="daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7" Dec 02 01:58:47 crc kubenswrapper[4884]: E1202 01:58:47.647773 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7\": container with ID starting with daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7 not found: ID does not exist" containerID="daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.647865 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7"} err="failed to get container status \"daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7\": rpc error: code = NotFound desc = could not find container \"daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7\": container with ID starting with daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7 not found: ID does not exist" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.647956 4884 scope.go:117] "RemoveContainer" containerID="5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d" Dec 02 01:58:47 crc kubenswrapper[4884]: E1202 01:58:47.648231 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d\": container with ID starting with 5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d not found: ID does not exist" containerID="5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.648257 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d"} err="failed to get container status \"5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d\": rpc error: code = NotFound desc = could not find container \"5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d\": container with ID starting with 5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d not found: ID does not exist" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.648272 4884 scope.go:117] "RemoveContainer" containerID="daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.648490 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7"} err="failed to get container status \"daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7\": rpc error: code = NotFound desc = could not find container \"daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7\": container with ID starting with daf5105f14c8fb720732d85ed545258f34f6830a15e86fea4da4d5dd15bc4de7 not found: ID does not exist" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.648510 4884 scope.go:117] "RemoveContainer" containerID="5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.648797 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d"} err="failed to get container status \"5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d\": rpc error: code = NotFound desc = could not find container \"5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d\": container with ID starting with 5ccf5c4a1d84b7bee20155b56964890db6201464b1d962f9522198da223e787d not found: ID does not exist" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.666386 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.666467 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.666541 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-scripts\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.666602 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-config-data\") pod \"nova-metadata-0\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.666642 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.666694 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-config-data\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.666723 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.666851 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.666922 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3de536b-47a0-4884-b6fb-3021b4ecbb98-run-httpd\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.666945 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89gfv\" (UniqueName: \"kubernetes.io/projected/d3de536b-47a0-4884-b6fb-3021b4ecbb98-kube-api-access-89gfv\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.667015 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg4x9\" (UniqueName: \"kubernetes.io/projected/062eea62-0f0f-45c4-a2e2-24820062d522-kube-api-access-vg4x9\") pod \"nova-metadata-0\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.667055 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3de536b-47a0-4884-b6fb-3021b4ecbb98-log-httpd\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.667129 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/062eea62-0f0f-45c4-a2e2-24820062d522-logs\") pod \"nova-metadata-0\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: E1202 01:58:47.723063 4884 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="38c77e578d9a2599b4508ff28f7a7b043df1e1b1369636bfdb22060b9f4f049d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 02 01:58:47 crc kubenswrapper[4884]: E1202 01:58:47.724362 4884 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="38c77e578d9a2599b4508ff28f7a7b043df1e1b1369636bfdb22060b9f4f049d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 02 01:58:47 crc kubenswrapper[4884]: E1202 01:58:47.726230 4884 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="38c77e578d9a2599b4508ff28f7a7b043df1e1b1369636bfdb22060b9f4f049d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 02 01:58:47 crc kubenswrapper[4884]: E1202 01:58:47.726271 4884 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="98757586-4157-403c-8a34-f7d28f2451cc" containerName="nova-scheduler-scheduler" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.768217 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg4x9\" (UniqueName: \"kubernetes.io/projected/062eea62-0f0f-45c4-a2e2-24820062d522-kube-api-access-vg4x9\") pod \"nova-metadata-0\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.768453 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3de536b-47a0-4884-b6fb-3021b4ecbb98-log-httpd\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.768595 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/062eea62-0f0f-45c4-a2e2-24820062d522-logs\") pod \"nova-metadata-0\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.768740 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.768867 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.768949 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-scripts\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.769047 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-config-data\") pod \"nova-metadata-0\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.769126 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.769203 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-config-data\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.769308 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.769388 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.769463 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3de536b-47a0-4884-b6fb-3021b4ecbb98-run-httpd\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.769528 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89gfv\" (UniqueName: \"kubernetes.io/projected/d3de536b-47a0-4884-b6fb-3021b4ecbb98-kube-api-access-89gfv\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.769226 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/062eea62-0f0f-45c4-a2e2-24820062d522-logs\") pod \"nova-metadata-0\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.768900 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3de536b-47a0-4884-b6fb-3021b4ecbb98-log-httpd\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.774377 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-scripts\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.774825 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.775049 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3de536b-47a0-4884-b6fb-3021b4ecbb98-run-httpd\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.775702 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.777204 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.778272 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.779060 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-config-data\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.783844 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-config-data\") pod \"nova-metadata-0\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.784763 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.785240 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg4x9\" (UniqueName: \"kubernetes.io/projected/062eea62-0f0f-45c4-a2e2-24820062d522-kube-api-access-vg4x9\") pod \"nova-metadata-0\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " pod="openstack/nova-metadata-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.787960 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89gfv\" (UniqueName: \"kubernetes.io/projected/d3de536b-47a0-4884-b6fb-3021b4ecbb98-kube-api-access-89gfv\") pod \"ceilometer-0\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.884707 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:58:47 crc kubenswrapper[4884]: I1202 01:58:47.892442 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 01:58:48 crc kubenswrapper[4884]: W1202 01:58:48.422918 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod062eea62_0f0f_45c4_a2e2_24820062d522.slice/crio-50e67ea73a7cfa5e4e71d3a9b63e955178eac72b584d3c64dd0155facae7532c WatchSource:0}: Error finding container 50e67ea73a7cfa5e4e71d3a9b63e955178eac72b584d3c64dd0155facae7532c: Status 404 returned error can't find the container with id 50e67ea73a7cfa5e4e71d3a9b63e955178eac72b584d3c64dd0155facae7532c Dec 02 01:58:48 crc kubenswrapper[4884]: I1202 01:58:48.426368 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:58:48 crc kubenswrapper[4884]: W1202 01:58:48.505822 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3de536b_47a0_4884_b6fb_3021b4ecbb98.slice/crio-2b278c38f70621e92e55bb2dea375dbf45c48c9a7fec8593d85c6672cc4f7ca6 WatchSource:0}: Error finding container 2b278c38f70621e92e55bb2dea375dbf45c48c9a7fec8593d85c6672cc4f7ca6: Status 404 returned error can't find the container with id 2b278c38f70621e92e55bb2dea375dbf45c48c9a7fec8593d85c6672cc4f7ca6 Dec 02 01:58:48 crc kubenswrapper[4884]: I1202 01:58:48.512586 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.433440 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.446070 4884 generic.go:334] "Generic (PLEG): container finished" podID="98757586-4157-403c-8a34-f7d28f2451cc" containerID="38c77e578d9a2599b4508ff28f7a7b043df1e1b1369636bfdb22060b9f4f049d" exitCode=0 Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.446167 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"98757586-4157-403c-8a34-f7d28f2451cc","Type":"ContainerDied","Data":"38c77e578d9a2599b4508ff28f7a7b043df1e1b1369636bfdb22060b9f4f049d"} Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.451406 4884 generic.go:334] "Generic (PLEG): container finished" podID="69686a6c-cdef-4fdf-8838-2624f841a85a" containerID="a7434306ef54ec5129024e57360f772e7cac65d9788835fe4193fe0ed268838b" exitCode=0 Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.451484 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"69686a6c-cdef-4fdf-8838-2624f841a85a","Type":"ContainerDied","Data":"a7434306ef54ec5129024e57360f772e7cac65d9788835fe4193fe0ed268838b"} Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.451517 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"69686a6c-cdef-4fdf-8838-2624f841a85a","Type":"ContainerDied","Data":"4188b9e7dcc7cdd426348da93f02fe6386d14ae7e6331066d7b529c037e78220"} Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.451537 4884 scope.go:117] "RemoveContainer" containerID="a7434306ef54ec5129024e57360f772e7cac65d9788835fe4193fe0ed268838b" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.451676 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.457899 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"062eea62-0f0f-45c4-a2e2-24820062d522","Type":"ContainerStarted","Data":"dcbd98665831ff5ed42a0debbf06d1ce2e3efeff9859ff024feeae70f3ac0952"} Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.458127 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"062eea62-0f0f-45c4-a2e2-24820062d522","Type":"ContainerStarted","Data":"5eba1d389242792d3f6698b0d5b8aa88025a5f1c27ab83416aa209c596ce5ba5"} Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.458138 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"062eea62-0f0f-45c4-a2e2-24820062d522","Type":"ContainerStarted","Data":"50e67ea73a7cfa5e4e71d3a9b63e955178eac72b584d3c64dd0155facae7532c"} Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.463875 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3de536b-47a0-4884-b6fb-3021b4ecbb98","Type":"ContainerStarted","Data":"59c87f346db8a4066aad1039e1b53d620c6bac171e09b11b40c5e62095f3c2d0"} Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.463955 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3de536b-47a0-4884-b6fb-3021b4ecbb98","Type":"ContainerStarted","Data":"2b278c38f70621e92e55bb2dea375dbf45c48c9a7fec8593d85c6672cc4f7ca6"} Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.485567 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.485546892 podStartE2EDuration="2.485546892s" podCreationTimestamp="2025-12-02 01:58:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:58:49.478499069 +0000 UTC m=+1346.154335983" watchObservedRunningTime="2025-12-02 01:58:49.485546892 +0000 UTC m=+1346.161383776" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.489638 4884 scope.go:117] "RemoveContainer" containerID="c1a1d3ff248af8626620469e1b48be5a6752079d2d7662c2414d0b2df7e2d98b" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.519684 4884 scope.go:117] "RemoveContainer" containerID="a7434306ef54ec5129024e57360f772e7cac65d9788835fe4193fe0ed268838b" Dec 02 01:58:49 crc kubenswrapper[4884]: E1202 01:58:49.520298 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7434306ef54ec5129024e57360f772e7cac65d9788835fe4193fe0ed268838b\": container with ID starting with a7434306ef54ec5129024e57360f772e7cac65d9788835fe4193fe0ed268838b not found: ID does not exist" containerID="a7434306ef54ec5129024e57360f772e7cac65d9788835fe4193fe0ed268838b" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.520457 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7434306ef54ec5129024e57360f772e7cac65d9788835fe4193fe0ed268838b"} err="failed to get container status \"a7434306ef54ec5129024e57360f772e7cac65d9788835fe4193fe0ed268838b\": rpc error: code = NotFound desc = could not find container \"a7434306ef54ec5129024e57360f772e7cac65d9788835fe4193fe0ed268838b\": container with ID starting with a7434306ef54ec5129024e57360f772e7cac65d9788835fe4193fe0ed268838b not found: ID does not exist" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.520641 4884 scope.go:117] "RemoveContainer" containerID="c1a1d3ff248af8626620469e1b48be5a6752079d2d7662c2414d0b2df7e2d98b" Dec 02 01:58:49 crc kubenswrapper[4884]: E1202 01:58:49.521967 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1a1d3ff248af8626620469e1b48be5a6752079d2d7662c2414d0b2df7e2d98b\": container with ID starting with c1a1d3ff248af8626620469e1b48be5a6752079d2d7662c2414d0b2df7e2d98b not found: ID does not exist" containerID="c1a1d3ff248af8626620469e1b48be5a6752079d2d7662c2414d0b2df7e2d98b" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.522018 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1a1d3ff248af8626620469e1b48be5a6752079d2d7662c2414d0b2df7e2d98b"} err="failed to get container status \"c1a1d3ff248af8626620469e1b48be5a6752079d2d7662c2414d0b2df7e2d98b\": rpc error: code = NotFound desc = could not find container \"c1a1d3ff248af8626620469e1b48be5a6752079d2d7662c2414d0b2df7e2d98b\": container with ID starting with c1a1d3ff248af8626620469e1b48be5a6752079d2d7662c2414d0b2df7e2d98b not found: ID does not exist" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.581240 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.604441 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69686a6c-cdef-4fdf-8838-2624f841a85a-logs\") pod \"69686a6c-cdef-4fdf-8838-2624f841a85a\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.604555 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69686a6c-cdef-4fdf-8838-2624f841a85a-combined-ca-bundle\") pod \"69686a6c-cdef-4fdf-8838-2624f841a85a\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.604670 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69686a6c-cdef-4fdf-8838-2624f841a85a-config-data\") pod \"69686a6c-cdef-4fdf-8838-2624f841a85a\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.604872 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tk6d\" (UniqueName: \"kubernetes.io/projected/69686a6c-cdef-4fdf-8838-2624f841a85a-kube-api-access-8tk6d\") pod \"69686a6c-cdef-4fdf-8838-2624f841a85a\" (UID: \"69686a6c-cdef-4fdf-8838-2624f841a85a\") " Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.605105 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69686a6c-cdef-4fdf-8838-2624f841a85a-logs" (OuterVolumeSpecName: "logs") pod "69686a6c-cdef-4fdf-8838-2624f841a85a" (UID: "69686a6c-cdef-4fdf-8838-2624f841a85a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.605514 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69686a6c-cdef-4fdf-8838-2624f841a85a-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.634935 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69686a6c-cdef-4fdf-8838-2624f841a85a-config-data" (OuterVolumeSpecName: "config-data") pod "69686a6c-cdef-4fdf-8838-2624f841a85a" (UID: "69686a6c-cdef-4fdf-8838-2624f841a85a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.637163 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69686a6c-cdef-4fdf-8838-2624f841a85a-kube-api-access-8tk6d" (OuterVolumeSpecName: "kube-api-access-8tk6d") pod "69686a6c-cdef-4fdf-8838-2624f841a85a" (UID: "69686a6c-cdef-4fdf-8838-2624f841a85a"). InnerVolumeSpecName "kube-api-access-8tk6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.644456 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69686a6c-cdef-4fdf-8838-2624f841a85a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69686a6c-cdef-4fdf-8838-2624f841a85a" (UID: "69686a6c-cdef-4fdf-8838-2624f841a85a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.706472 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzvf4\" (UniqueName: \"kubernetes.io/projected/98757586-4157-403c-8a34-f7d28f2451cc-kube-api-access-rzvf4\") pod \"98757586-4157-403c-8a34-f7d28f2451cc\" (UID: \"98757586-4157-403c-8a34-f7d28f2451cc\") " Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.707293 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98757586-4157-403c-8a34-f7d28f2451cc-config-data\") pod \"98757586-4157-403c-8a34-f7d28f2451cc\" (UID: \"98757586-4157-403c-8a34-f7d28f2451cc\") " Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.707900 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98757586-4157-403c-8a34-f7d28f2451cc-combined-ca-bundle\") pod \"98757586-4157-403c-8a34-f7d28f2451cc\" (UID: \"98757586-4157-403c-8a34-f7d28f2451cc\") " Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.709719 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tk6d\" (UniqueName: \"kubernetes.io/projected/69686a6c-cdef-4fdf-8838-2624f841a85a-kube-api-access-8tk6d\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.709773 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69686a6c-cdef-4fdf-8838-2624f841a85a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.709786 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69686a6c-cdef-4fdf-8838-2624f841a85a-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.712064 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98757586-4157-403c-8a34-f7d28f2451cc-kube-api-access-rzvf4" (OuterVolumeSpecName: "kube-api-access-rzvf4") pod "98757586-4157-403c-8a34-f7d28f2451cc" (UID: "98757586-4157-403c-8a34-f7d28f2451cc"). InnerVolumeSpecName "kube-api-access-rzvf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.740147 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98757586-4157-403c-8a34-f7d28f2451cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98757586-4157-403c-8a34-f7d28f2451cc" (UID: "98757586-4157-403c-8a34-f7d28f2451cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.740575 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98757586-4157-403c-8a34-f7d28f2451cc-config-data" (OuterVolumeSpecName: "config-data") pod "98757586-4157-403c-8a34-f7d28f2451cc" (UID: "98757586-4157-403c-8a34-f7d28f2451cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.811010 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzvf4\" (UniqueName: \"kubernetes.io/projected/98757586-4157-403c-8a34-f7d28f2451cc-kube-api-access-rzvf4\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.811039 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98757586-4157-403c-8a34-f7d28f2451cc-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.811053 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98757586-4157-403c-8a34-f7d28f2451cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.838987 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.847390 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.868919 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 02 01:58:49 crc kubenswrapper[4884]: E1202 01:58:49.869254 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69686a6c-cdef-4fdf-8838-2624f841a85a" containerName="nova-api-log" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.869265 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="69686a6c-cdef-4fdf-8838-2624f841a85a" containerName="nova-api-log" Dec 02 01:58:49 crc kubenswrapper[4884]: E1202 01:58:49.869283 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98757586-4157-403c-8a34-f7d28f2451cc" containerName="nova-scheduler-scheduler" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.869290 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="98757586-4157-403c-8a34-f7d28f2451cc" containerName="nova-scheduler-scheduler" Dec 02 01:58:49 crc kubenswrapper[4884]: E1202 01:58:49.869303 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69686a6c-cdef-4fdf-8838-2624f841a85a" containerName="nova-api-api" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.869309 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="69686a6c-cdef-4fdf-8838-2624f841a85a" containerName="nova-api-api" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.869483 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="98757586-4157-403c-8a34-f7d28f2451cc" containerName="nova-scheduler-scheduler" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.869497 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="69686a6c-cdef-4fdf-8838-2624f841a85a" containerName="nova-api-api" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.869512 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="69686a6c-cdef-4fdf-8838-2624f841a85a" containerName="nova-api-log" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.870432 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.870502 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 01:58:49 crc kubenswrapper[4884]: I1202 01:58:49.908559 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.015177 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaf83795-11a7-477a-9ecb-a2659ce4deda-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " pod="openstack/nova-api-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.015293 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf83795-11a7-477a-9ecb-a2659ce4deda-config-data\") pod \"nova-api-0\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " pod="openstack/nova-api-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.015428 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lct6l\" (UniqueName: \"kubernetes.io/projected/eaf83795-11a7-477a-9ecb-a2659ce4deda-kube-api-access-lct6l\") pod \"nova-api-0\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " pod="openstack/nova-api-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.015609 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaf83795-11a7-477a-9ecb-a2659ce4deda-logs\") pod \"nova-api-0\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " pod="openstack/nova-api-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.119205 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaf83795-11a7-477a-9ecb-a2659ce4deda-logs\") pod \"nova-api-0\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " pod="openstack/nova-api-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.119356 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaf83795-11a7-477a-9ecb-a2659ce4deda-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " pod="openstack/nova-api-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.119392 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf83795-11a7-477a-9ecb-a2659ce4deda-config-data\") pod \"nova-api-0\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " pod="openstack/nova-api-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.119417 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lct6l\" (UniqueName: \"kubernetes.io/projected/eaf83795-11a7-477a-9ecb-a2659ce4deda-kube-api-access-lct6l\") pod \"nova-api-0\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " pod="openstack/nova-api-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.119643 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaf83795-11a7-477a-9ecb-a2659ce4deda-logs\") pod \"nova-api-0\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " pod="openstack/nova-api-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.123310 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaf83795-11a7-477a-9ecb-a2659ce4deda-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " pod="openstack/nova-api-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.123592 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf83795-11a7-477a-9ecb-a2659ce4deda-config-data\") pod \"nova-api-0\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " pod="openstack/nova-api-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.140328 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lct6l\" (UniqueName: \"kubernetes.io/projected/eaf83795-11a7-477a-9ecb-a2659ce4deda-kube-api-access-lct6l\") pod \"nova-api-0\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " pod="openstack/nova-api-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.229841 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.480797 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"98757586-4157-403c-8a34-f7d28f2451cc","Type":"ContainerDied","Data":"3b7b070da3fa28f84f8934e769745bd930e042b6a030252dc58e38db24afab6d"} Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.481177 4884 scope.go:117] "RemoveContainer" containerID="38c77e578d9a2599b4508ff28f7a7b043df1e1b1369636bfdb22060b9f4f049d" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.480842 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.520796 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.575426 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.588014 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.589665 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.591961 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.596223 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.673953 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.683772 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17a0f0d-1e28-4238-83f8-742b3b3df02c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d17a0f0d-1e28-4238-83f8-742b3b3df02c\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.683934 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17a0f0d-1e28-4238-83f8-742b3b3df02c-config-data\") pod \"nova-scheduler-0\" (UID: \"d17a0f0d-1e28-4238-83f8-742b3b3df02c\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.683971 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5rr8\" (UniqueName: \"kubernetes.io/projected/d17a0f0d-1e28-4238-83f8-742b3b3df02c-kube-api-access-d5rr8\") pod \"nova-scheduler-0\" (UID: \"d17a0f0d-1e28-4238-83f8-742b3b3df02c\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.786016 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17a0f0d-1e28-4238-83f8-742b3b3df02c-config-data\") pod \"nova-scheduler-0\" (UID: \"d17a0f0d-1e28-4238-83f8-742b3b3df02c\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.786068 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5rr8\" (UniqueName: \"kubernetes.io/projected/d17a0f0d-1e28-4238-83f8-742b3b3df02c-kube-api-access-d5rr8\") pod \"nova-scheduler-0\" (UID: \"d17a0f0d-1e28-4238-83f8-742b3b3df02c\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.786225 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17a0f0d-1e28-4238-83f8-742b3b3df02c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d17a0f0d-1e28-4238-83f8-742b3b3df02c\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.792441 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17a0f0d-1e28-4238-83f8-742b3b3df02c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d17a0f0d-1e28-4238-83f8-742b3b3df02c\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.796941 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17a0f0d-1e28-4238-83f8-742b3b3df02c-config-data\") pod \"nova-scheduler-0\" (UID: \"d17a0f0d-1e28-4238-83f8-742b3b3df02c\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.805984 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5rr8\" (UniqueName: \"kubernetes.io/projected/d17a0f0d-1e28-4238-83f8-742b3b3df02c-kube-api-access-d5rr8\") pod \"nova-scheduler-0\" (UID: \"d17a0f0d-1e28-4238-83f8-742b3b3df02c\") " pod="openstack/nova-scheduler-0" Dec 02 01:58:50 crc kubenswrapper[4884]: I1202 01:58:50.906811 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 01:58:51 crc kubenswrapper[4884]: I1202 01:58:51.403275 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 01:58:51 crc kubenswrapper[4884]: W1202 01:58:51.453995 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd17a0f0d_1e28_4238_83f8_742b3b3df02c.slice/crio-a5727aa6caf1ac333652dbbee5fc3dcf4b7a5dd14b0d84dfcc590576198bbf3f WatchSource:0}: Error finding container a5727aa6caf1ac333652dbbee5fc3dcf4b7a5dd14b0d84dfcc590576198bbf3f: Status 404 returned error can't find the container with id a5727aa6caf1ac333652dbbee5fc3dcf4b7a5dd14b0d84dfcc590576198bbf3f Dec 02 01:58:51 crc kubenswrapper[4884]: I1202 01:58:51.492375 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3de536b-47a0-4884-b6fb-3021b4ecbb98","Type":"ContainerStarted","Data":"d83ec64874b2bf694026e9c2b12dd39ab9de100fdb6207cde5e15afad02cab9f"} Dec 02 01:58:51 crc kubenswrapper[4884]: I1202 01:58:51.493388 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eaf83795-11a7-477a-9ecb-a2659ce4deda","Type":"ContainerStarted","Data":"be1699158e9ea560fdd25573e7c4baab0c870736e5eb55256359d82f90050ef6"} Dec 02 01:58:51 crc kubenswrapper[4884]: I1202 01:58:51.493406 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eaf83795-11a7-477a-9ecb-a2659ce4deda","Type":"ContainerStarted","Data":"1de553a8de08fd0a8f3b7c4b482318858805412993c85abf07293413a9613402"} Dec 02 01:58:51 crc kubenswrapper[4884]: I1202 01:58:51.498624 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d17a0f0d-1e28-4238-83f8-742b3b3df02c","Type":"ContainerStarted","Data":"a5727aa6caf1ac333652dbbee5fc3dcf4b7a5dd14b0d84dfcc590576198bbf3f"} Dec 02 01:58:51 crc kubenswrapper[4884]: I1202 01:58:51.629120 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69686a6c-cdef-4fdf-8838-2624f841a85a" path="/var/lib/kubelet/pods/69686a6c-cdef-4fdf-8838-2624f841a85a/volumes" Dec 02 01:58:51 crc kubenswrapper[4884]: I1202 01:58:51.629902 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98757586-4157-403c-8a34-f7d28f2451cc" path="/var/lib/kubelet/pods/98757586-4157-403c-8a34-f7d28f2451cc/volumes" Dec 02 01:58:52 crc kubenswrapper[4884]: I1202 01:58:52.522150 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eaf83795-11a7-477a-9ecb-a2659ce4deda","Type":"ContainerStarted","Data":"a334be3738065008f1cc04000ba8e76c1c0cf8fc7b89df9cc376d227b0071fff"} Dec 02 01:58:52 crc kubenswrapper[4884]: I1202 01:58:52.527278 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d17a0f0d-1e28-4238-83f8-742b3b3df02c","Type":"ContainerStarted","Data":"6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd"} Dec 02 01:58:52 crc kubenswrapper[4884]: I1202 01:58:52.530153 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3de536b-47a0-4884-b6fb-3021b4ecbb98","Type":"ContainerStarted","Data":"56dc8361d659338a5ec8ff240b527700ca795747e611ad29313581993b3fe876"} Dec 02 01:58:52 crc kubenswrapper[4884]: I1202 01:58:52.549656 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.549629526 podStartE2EDuration="3.549629526s" podCreationTimestamp="2025-12-02 01:58:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:58:52.547152225 +0000 UTC m=+1349.222989199" watchObservedRunningTime="2025-12-02 01:58:52.549629526 +0000 UTC m=+1349.225466410" Dec 02 01:58:52 crc kubenswrapper[4884]: I1202 01:58:52.578711 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.578690842 podStartE2EDuration="2.578690842s" podCreationTimestamp="2025-12-02 01:58:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:58:52.567912786 +0000 UTC m=+1349.243749670" watchObservedRunningTime="2025-12-02 01:58:52.578690842 +0000 UTC m=+1349.254527726" Dec 02 01:58:52 crc kubenswrapper[4884]: I1202 01:58:52.893489 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 01:58:52 crc kubenswrapper[4884]: I1202 01:58:52.894866 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 01:58:53 crc kubenswrapper[4884]: I1202 01:58:53.544615 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3de536b-47a0-4884-b6fb-3021b4ecbb98","Type":"ContainerStarted","Data":"d7821da0426d4142c952cc46bc599f50b094801e95a7e2c96baa1a1f215e33cf"} Dec 02 01:58:53 crc kubenswrapper[4884]: I1202 01:58:53.545265 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 01:58:53 crc kubenswrapper[4884]: I1202 01:58:53.586755 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.049672565 podStartE2EDuration="6.586722421s" podCreationTimestamp="2025-12-02 01:58:47 +0000 UTC" firstStartedPulling="2025-12-02 01:58:48.509168962 +0000 UTC m=+1345.185005846" lastFinishedPulling="2025-12-02 01:58:53.046218778 +0000 UTC m=+1349.722055702" observedRunningTime="2025-12-02 01:58:53.578064658 +0000 UTC m=+1350.253901602" watchObservedRunningTime="2025-12-02 01:58:53.586722421 +0000 UTC m=+1350.262559305" Dec 02 01:58:53 crc kubenswrapper[4884]: I1202 01:58:53.726547 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 02 01:58:54 crc kubenswrapper[4884]: I1202 01:58:54.805903 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 02 01:58:55 crc kubenswrapper[4884]: I1202 01:58:55.908180 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 02 01:58:57 crc kubenswrapper[4884]: I1202 01:58:57.893532 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 02 01:58:57 crc kubenswrapper[4884]: I1202 01:58:57.894673 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 02 01:58:58 crc kubenswrapper[4884]: I1202 01:58:58.909905 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="062eea62-0f0f-45c4-a2e2-24820062d522" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 01:58:58 crc kubenswrapper[4884]: I1202 01:58:58.909975 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="062eea62-0f0f-45c4-a2e2-24820062d522" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 01:59:00 crc kubenswrapper[4884]: I1202 01:59:00.230501 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 01:59:00 crc kubenswrapper[4884]: I1202 01:59:00.230601 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 01:59:00 crc kubenswrapper[4884]: I1202 01:59:00.908366 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 02 01:59:00 crc kubenswrapper[4884]: I1202 01:59:00.945937 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 02 01:59:01 crc kubenswrapper[4884]: I1202 01:59:01.271983 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="eaf83795-11a7-477a-9ecb-a2659ce4deda" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.212:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 01:59:01 crc kubenswrapper[4884]: I1202 01:59:01.312949 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="eaf83795-11a7-477a-9ecb-a2659ce4deda" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.212:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 01:59:01 crc kubenswrapper[4884]: I1202 01:59:01.674493 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 02 01:59:07 crc kubenswrapper[4884]: I1202 01:59:07.900897 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 02 01:59:07 crc kubenswrapper[4884]: I1202 01:59:07.903052 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 02 01:59:07 crc kubenswrapper[4884]: I1202 01:59:07.910514 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 02 01:59:08 crc kubenswrapper[4884]: I1202 01:59:08.757182 4884 generic.go:334] "Generic (PLEG): container finished" podID="bf6e71cc-1f65-462c-8234-3c93203d1728" containerID="bf92762a1db3ee254be4c737489b66d5ec6b31d47378a862216842b18ac20a7a" exitCode=137 Dec 02 01:59:08 crc kubenswrapper[4884]: I1202 01:59:08.757253 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bf6e71cc-1f65-462c-8234-3c93203d1728","Type":"ContainerDied","Data":"bf92762a1db3ee254be4c737489b66d5ec6b31d47378a862216842b18ac20a7a"} Dec 02 01:59:08 crc kubenswrapper[4884]: I1202 01:59:08.758004 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bf6e71cc-1f65-462c-8234-3c93203d1728","Type":"ContainerDied","Data":"d5708b5383466dc627b5aed79db3e94818e0a8ba2546ffe7370f4afce88345e4"} Dec 02 01:59:08 crc kubenswrapper[4884]: I1202 01:59:08.758043 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5708b5383466dc627b5aed79db3e94818e0a8ba2546ffe7370f4afce88345e4" Dec 02 01:59:08 crc kubenswrapper[4884]: I1202 01:59:08.763821 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 02 01:59:08 crc kubenswrapper[4884]: I1202 01:59:08.782326 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:08 crc kubenswrapper[4884]: I1202 01:59:08.897725 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5wsq\" (UniqueName: \"kubernetes.io/projected/bf6e71cc-1f65-462c-8234-3c93203d1728-kube-api-access-f5wsq\") pod \"bf6e71cc-1f65-462c-8234-3c93203d1728\" (UID: \"bf6e71cc-1f65-462c-8234-3c93203d1728\") " Dec 02 01:59:08 crc kubenswrapper[4884]: I1202 01:59:08.898070 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf6e71cc-1f65-462c-8234-3c93203d1728-combined-ca-bundle\") pod \"bf6e71cc-1f65-462c-8234-3c93203d1728\" (UID: \"bf6e71cc-1f65-462c-8234-3c93203d1728\") " Dec 02 01:59:08 crc kubenswrapper[4884]: I1202 01:59:08.898186 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf6e71cc-1f65-462c-8234-3c93203d1728-config-data\") pod \"bf6e71cc-1f65-462c-8234-3c93203d1728\" (UID: \"bf6e71cc-1f65-462c-8234-3c93203d1728\") " Dec 02 01:59:08 crc kubenswrapper[4884]: I1202 01:59:08.903108 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf6e71cc-1f65-462c-8234-3c93203d1728-kube-api-access-f5wsq" (OuterVolumeSpecName: "kube-api-access-f5wsq") pod "bf6e71cc-1f65-462c-8234-3c93203d1728" (UID: "bf6e71cc-1f65-462c-8234-3c93203d1728"). InnerVolumeSpecName "kube-api-access-f5wsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:59:08 crc kubenswrapper[4884]: I1202 01:59:08.931511 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf6e71cc-1f65-462c-8234-3c93203d1728-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf6e71cc-1f65-462c-8234-3c93203d1728" (UID: "bf6e71cc-1f65-462c-8234-3c93203d1728"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:08 crc kubenswrapper[4884]: I1202 01:59:08.950383 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf6e71cc-1f65-462c-8234-3c93203d1728-config-data" (OuterVolumeSpecName: "config-data") pod "bf6e71cc-1f65-462c-8234-3c93203d1728" (UID: "bf6e71cc-1f65-462c-8234-3c93203d1728"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:09 crc kubenswrapper[4884]: I1202 01:59:09.001330 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf6e71cc-1f65-462c-8234-3c93203d1728-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:09 crc kubenswrapper[4884]: I1202 01:59:09.001364 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf6e71cc-1f65-462c-8234-3c93203d1728-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:09 crc kubenswrapper[4884]: I1202 01:59:09.001374 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5wsq\" (UniqueName: \"kubernetes.io/projected/bf6e71cc-1f65-462c-8234-3c93203d1728-kube-api-access-f5wsq\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:09 crc kubenswrapper[4884]: I1202 01:59:09.768157 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:09 crc kubenswrapper[4884]: I1202 01:59:09.804980 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 01:59:09 crc kubenswrapper[4884]: I1202 01:59:09.824082 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 01:59:09 crc kubenswrapper[4884]: I1202 01:59:09.841419 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 01:59:09 crc kubenswrapper[4884]: E1202 01:59:09.842194 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6e71cc-1f65-462c-8234-3c93203d1728" containerName="nova-cell1-novncproxy-novncproxy" Dec 02 01:59:09 crc kubenswrapper[4884]: I1202 01:59:09.842225 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6e71cc-1f65-462c-8234-3c93203d1728" containerName="nova-cell1-novncproxy-novncproxy" Dec 02 01:59:09 crc kubenswrapper[4884]: I1202 01:59:09.842628 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf6e71cc-1f65-462c-8234-3c93203d1728" containerName="nova-cell1-novncproxy-novncproxy" Dec 02 01:59:09 crc kubenswrapper[4884]: I1202 01:59:09.843773 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:09 crc kubenswrapper[4884]: I1202 01:59:09.846956 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 02 01:59:09 crc kubenswrapper[4884]: I1202 01:59:09.847029 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 02 01:59:09 crc kubenswrapper[4884]: I1202 01:59:09.847160 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 02 01:59:09 crc kubenswrapper[4884]: I1202 01:59:09.852237 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.023331 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c72fc8df-eefd-43aa-879f-ac397417aa90-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c72fc8df-eefd-43aa-879f-ac397417aa90\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.023579 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c72fc8df-eefd-43aa-879f-ac397417aa90-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c72fc8df-eefd-43aa-879f-ac397417aa90\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.024041 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c72fc8df-eefd-43aa-879f-ac397417aa90-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c72fc8df-eefd-43aa-879f-ac397417aa90\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.024115 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkgr4\" (UniqueName: \"kubernetes.io/projected/c72fc8df-eefd-43aa-879f-ac397417aa90-kube-api-access-fkgr4\") pod \"nova-cell1-novncproxy-0\" (UID: \"c72fc8df-eefd-43aa-879f-ac397417aa90\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.024185 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c72fc8df-eefd-43aa-879f-ac397417aa90-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c72fc8df-eefd-43aa-879f-ac397417aa90\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.126410 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c72fc8df-eefd-43aa-879f-ac397417aa90-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c72fc8df-eefd-43aa-879f-ac397417aa90\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.126620 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c72fc8df-eefd-43aa-879f-ac397417aa90-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c72fc8df-eefd-43aa-879f-ac397417aa90\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.126790 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c72fc8df-eefd-43aa-879f-ac397417aa90-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c72fc8df-eefd-43aa-879f-ac397417aa90\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.126838 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkgr4\" (UniqueName: \"kubernetes.io/projected/c72fc8df-eefd-43aa-879f-ac397417aa90-kube-api-access-fkgr4\") pod \"nova-cell1-novncproxy-0\" (UID: \"c72fc8df-eefd-43aa-879f-ac397417aa90\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.126877 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c72fc8df-eefd-43aa-879f-ac397417aa90-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c72fc8df-eefd-43aa-879f-ac397417aa90\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.131851 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/c72fc8df-eefd-43aa-879f-ac397417aa90-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c72fc8df-eefd-43aa-879f-ac397417aa90\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.139885 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c72fc8df-eefd-43aa-879f-ac397417aa90-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"c72fc8df-eefd-43aa-879f-ac397417aa90\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.139973 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c72fc8df-eefd-43aa-879f-ac397417aa90-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"c72fc8df-eefd-43aa-879f-ac397417aa90\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.140246 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/c72fc8df-eefd-43aa-879f-ac397417aa90-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"c72fc8df-eefd-43aa-879f-ac397417aa90\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.143337 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkgr4\" (UniqueName: \"kubernetes.io/projected/c72fc8df-eefd-43aa-879f-ac397417aa90-kube-api-access-fkgr4\") pod \"nova-cell1-novncproxy-0\" (UID: \"c72fc8df-eefd-43aa-879f-ac397417aa90\") " pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.183083 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.234946 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.235063 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.235493 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.235565 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.240541 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.241642 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.456453 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-j6jlj"] Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.459613 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.490851 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-j6jlj"] Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.540532 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-config\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.540621 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgjm9\" (UniqueName: \"kubernetes.io/projected/3e49c865-898a-496f-80b5-32c3ed00558b-kube-api-access-fgjm9\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.540649 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.540797 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.541008 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.541037 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.642668 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-config\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.642788 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgjm9\" (UniqueName: \"kubernetes.io/projected/3e49c865-898a-496f-80b5-32c3ed00558b-kube-api-access-fgjm9\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.642818 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.642859 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.642907 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.642926 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.643951 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.644525 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-config\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.645135 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.648290 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.648318 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.671094 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgjm9\" (UniqueName: \"kubernetes.io/projected/3e49c865-898a-496f-80b5-32c3ed00558b-kube-api-access-fgjm9\") pod \"dnsmasq-dns-89c5cd4d5-j6jlj\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.776607 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 02 01:59:10 crc kubenswrapper[4884]: I1202 01:59:10.789255 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:10 crc kubenswrapper[4884]: W1202 01:59:10.798973 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc72fc8df_eefd_43aa_879f_ac397417aa90.slice/crio-858afede7793b01a71c3d7909e76acb93ab424919952542875b212053a6f632d WatchSource:0}: Error finding container 858afede7793b01a71c3d7909e76acb93ab424919952542875b212053a6f632d: Status 404 returned error can't find the container with id 858afede7793b01a71c3d7909e76acb93ab424919952542875b212053a6f632d Dec 02 01:59:11 crc kubenswrapper[4884]: I1202 01:59:11.326241 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-j6jlj"] Dec 02 01:59:11 crc kubenswrapper[4884]: W1202 01:59:11.331768 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e49c865_898a_496f_80b5_32c3ed00558b.slice/crio-3407bfc3b3f882b7feff4a4a15c77e0700bbffb70a44374e80fbaa3cd77cf341 WatchSource:0}: Error finding container 3407bfc3b3f882b7feff4a4a15c77e0700bbffb70a44374e80fbaa3cd77cf341: Status 404 returned error can't find the container with id 3407bfc3b3f882b7feff4a4a15c77e0700bbffb70a44374e80fbaa3cd77cf341 Dec 02 01:59:11 crc kubenswrapper[4884]: I1202 01:59:11.627426 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf6e71cc-1f65-462c-8234-3c93203d1728" path="/var/lib/kubelet/pods/bf6e71cc-1f65-462c-8234-3c93203d1728/volumes" Dec 02 01:59:11 crc kubenswrapper[4884]: I1202 01:59:11.799309 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c72fc8df-eefd-43aa-879f-ac397417aa90","Type":"ContainerStarted","Data":"1601355e2b963fc9e6977666f82727f236d08ffd6dfc107cbab9a431e7e67ff8"} Dec 02 01:59:11 crc kubenswrapper[4884]: I1202 01:59:11.799548 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"c72fc8df-eefd-43aa-879f-ac397417aa90","Type":"ContainerStarted","Data":"858afede7793b01a71c3d7909e76acb93ab424919952542875b212053a6f632d"} Dec 02 01:59:11 crc kubenswrapper[4884]: I1202 01:59:11.805572 4884 generic.go:334] "Generic (PLEG): container finished" podID="3e49c865-898a-496f-80b5-32c3ed00558b" containerID="54973ed23c0576f50ec34ce1ce3bc9ea2757bd5f4b3e2678e8c8129033cb1f7f" exitCode=0 Dec 02 01:59:11 crc kubenswrapper[4884]: I1202 01:59:11.805669 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" event={"ID":"3e49c865-898a-496f-80b5-32c3ed00558b","Type":"ContainerDied","Data":"54973ed23c0576f50ec34ce1ce3bc9ea2757bd5f4b3e2678e8c8129033cb1f7f"} Dec 02 01:59:11 crc kubenswrapper[4884]: I1202 01:59:11.805721 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" event={"ID":"3e49c865-898a-496f-80b5-32c3ed00558b","Type":"ContainerStarted","Data":"3407bfc3b3f882b7feff4a4a15c77e0700bbffb70a44374e80fbaa3cd77cf341"} Dec 02 01:59:11 crc kubenswrapper[4884]: I1202 01:59:11.826018 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.825999541 podStartE2EDuration="2.825999541s" podCreationTimestamp="2025-12-02 01:59:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:59:11.814300363 +0000 UTC m=+1368.490137267" watchObservedRunningTime="2025-12-02 01:59:11.825999541 +0000 UTC m=+1368.501836425" Dec 02 01:59:12 crc kubenswrapper[4884]: I1202 01:59:12.768330 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:59:12 crc kubenswrapper[4884]: I1202 01:59:12.768837 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="ceilometer-central-agent" containerID="cri-o://59c87f346db8a4066aad1039e1b53d620c6bac171e09b11b40c5e62095f3c2d0" gracePeriod=30 Dec 02 01:59:12 crc kubenswrapper[4884]: I1202 01:59:12.768939 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="proxy-httpd" containerID="cri-o://d7821da0426d4142c952cc46bc599f50b094801e95a7e2c96baa1a1f215e33cf" gracePeriod=30 Dec 02 01:59:12 crc kubenswrapper[4884]: I1202 01:59:12.768975 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="sg-core" containerID="cri-o://56dc8361d659338a5ec8ff240b527700ca795747e611ad29313581993b3fe876" gracePeriod=30 Dec 02 01:59:12 crc kubenswrapper[4884]: I1202 01:59:12.769233 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="ceilometer-notification-agent" containerID="cri-o://d83ec64874b2bf694026e9c2b12dd39ab9de100fdb6207cde5e15afad02cab9f" gracePeriod=30 Dec 02 01:59:12 crc kubenswrapper[4884]: I1202 01:59:12.785499 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.210:3000/\": EOF" Dec 02 01:59:12 crc kubenswrapper[4884]: I1202 01:59:12.816318 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" event={"ID":"3e49c865-898a-496f-80b5-32c3ed00558b","Type":"ContainerStarted","Data":"ca6e36d901f79f15f193e08db6cec23b2ac6ad6aa717ed56fb27891f1f22b341"} Dec 02 01:59:12 crc kubenswrapper[4884]: I1202 01:59:12.816508 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:12 crc kubenswrapper[4884]: I1202 01:59:12.837192 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" podStartSLOduration=2.837172759 podStartE2EDuration="2.837172759s" podCreationTimestamp="2025-12-02 01:59:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:59:12.833899798 +0000 UTC m=+1369.509736692" watchObservedRunningTime="2025-12-02 01:59:12.837172759 +0000 UTC m=+1369.513009633" Dec 02 01:59:13 crc kubenswrapper[4884]: I1202 01:59:13.000711 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:59:13 crc kubenswrapper[4884]: I1202 01:59:13.001353 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="eaf83795-11a7-477a-9ecb-a2659ce4deda" containerName="nova-api-api" containerID="cri-o://a334be3738065008f1cc04000ba8e76c1c0cf8fc7b89df9cc376d227b0071fff" gracePeriod=30 Dec 02 01:59:13 crc kubenswrapper[4884]: I1202 01:59:13.001302 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="eaf83795-11a7-477a-9ecb-a2659ce4deda" containerName="nova-api-log" containerID="cri-o://be1699158e9ea560fdd25573e7c4baab0c870736e5eb55256359d82f90050ef6" gracePeriod=30 Dec 02 01:59:13 crc kubenswrapper[4884]: I1202 01:59:13.829474 4884 generic.go:334] "Generic (PLEG): container finished" podID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerID="d7821da0426d4142c952cc46bc599f50b094801e95a7e2c96baa1a1f215e33cf" exitCode=0 Dec 02 01:59:13 crc kubenswrapper[4884]: I1202 01:59:13.829513 4884 generic.go:334] "Generic (PLEG): container finished" podID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerID="56dc8361d659338a5ec8ff240b527700ca795747e611ad29313581993b3fe876" exitCode=2 Dec 02 01:59:13 crc kubenswrapper[4884]: I1202 01:59:13.829524 4884 generic.go:334] "Generic (PLEG): container finished" podID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerID="59c87f346db8a4066aad1039e1b53d620c6bac171e09b11b40c5e62095f3c2d0" exitCode=0 Dec 02 01:59:13 crc kubenswrapper[4884]: I1202 01:59:13.829534 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3de536b-47a0-4884-b6fb-3021b4ecbb98","Type":"ContainerDied","Data":"d7821da0426d4142c952cc46bc599f50b094801e95a7e2c96baa1a1f215e33cf"} Dec 02 01:59:13 crc kubenswrapper[4884]: I1202 01:59:13.829574 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3de536b-47a0-4884-b6fb-3021b4ecbb98","Type":"ContainerDied","Data":"56dc8361d659338a5ec8ff240b527700ca795747e611ad29313581993b3fe876"} Dec 02 01:59:13 crc kubenswrapper[4884]: I1202 01:59:13.829587 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3de536b-47a0-4884-b6fb-3021b4ecbb98","Type":"ContainerDied","Data":"59c87f346db8a4066aad1039e1b53d620c6bac171e09b11b40c5e62095f3c2d0"} Dec 02 01:59:13 crc kubenswrapper[4884]: I1202 01:59:13.833091 4884 generic.go:334] "Generic (PLEG): container finished" podID="eaf83795-11a7-477a-9ecb-a2659ce4deda" containerID="be1699158e9ea560fdd25573e7c4baab0c870736e5eb55256359d82f90050ef6" exitCode=143 Dec 02 01:59:13 crc kubenswrapper[4884]: I1202 01:59:13.833227 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eaf83795-11a7-477a-9ecb-a2659ce4deda","Type":"ContainerDied","Data":"be1699158e9ea560fdd25573e7c4baab0c870736e5eb55256359d82f90050ef6"} Dec 02 01:59:15 crc kubenswrapper[4884]: I1202 01:59:15.184132 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.583943 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.587440 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.657521 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaf83795-11a7-477a-9ecb-a2659ce4deda-combined-ca-bundle\") pod \"eaf83795-11a7-477a-9ecb-a2659ce4deda\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.657622 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3de536b-47a0-4884-b6fb-3021b4ecbb98-run-httpd\") pod \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.657706 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf83795-11a7-477a-9ecb-a2659ce4deda-config-data\") pod \"eaf83795-11a7-477a-9ecb-a2659ce4deda\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.657795 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lct6l\" (UniqueName: \"kubernetes.io/projected/eaf83795-11a7-477a-9ecb-a2659ce4deda-kube-api-access-lct6l\") pod \"eaf83795-11a7-477a-9ecb-a2659ce4deda\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.657912 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-combined-ca-bundle\") pod \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.658022 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-config-data\") pod \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.658361 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-ceilometer-tls-certs\") pod \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.658454 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3de536b-47a0-4884-b6fb-3021b4ecbb98-log-httpd\") pod \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.658595 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3de536b-47a0-4884-b6fb-3021b4ecbb98-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d3de536b-47a0-4884-b6fb-3021b4ecbb98" (UID: "d3de536b-47a0-4884-b6fb-3021b4ecbb98"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.659026 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-scripts\") pod \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.659067 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89gfv\" (UniqueName: \"kubernetes.io/projected/d3de536b-47a0-4884-b6fb-3021b4ecbb98-kube-api-access-89gfv\") pod \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.659106 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaf83795-11a7-477a-9ecb-a2659ce4deda-logs\") pod \"eaf83795-11a7-477a-9ecb-a2659ce4deda\" (UID: \"eaf83795-11a7-477a-9ecb-a2659ce4deda\") " Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.659148 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-sg-core-conf-yaml\") pod \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\" (UID: \"d3de536b-47a0-4884-b6fb-3021b4ecbb98\") " Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.659210 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3de536b-47a0-4884-b6fb-3021b4ecbb98-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d3de536b-47a0-4884-b6fb-3021b4ecbb98" (UID: "d3de536b-47a0-4884-b6fb-3021b4ecbb98"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.659910 4884 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3de536b-47a0-4884-b6fb-3021b4ecbb98-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.659939 4884 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3de536b-47a0-4884-b6fb-3021b4ecbb98-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.660148 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eaf83795-11a7-477a-9ecb-a2659ce4deda-logs" (OuterVolumeSpecName: "logs") pod "eaf83795-11a7-477a-9ecb-a2659ce4deda" (UID: "eaf83795-11a7-477a-9ecb-a2659ce4deda"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.664637 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaf83795-11a7-477a-9ecb-a2659ce4deda-kube-api-access-lct6l" (OuterVolumeSpecName: "kube-api-access-lct6l") pod "eaf83795-11a7-477a-9ecb-a2659ce4deda" (UID: "eaf83795-11a7-477a-9ecb-a2659ce4deda"). InnerVolumeSpecName "kube-api-access-lct6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.672968 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3de536b-47a0-4884-b6fb-3021b4ecbb98-kube-api-access-89gfv" (OuterVolumeSpecName: "kube-api-access-89gfv") pod "d3de536b-47a0-4884-b6fb-3021b4ecbb98" (UID: "d3de536b-47a0-4884-b6fb-3021b4ecbb98"). InnerVolumeSpecName "kube-api-access-89gfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.673930 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-scripts" (OuterVolumeSpecName: "scripts") pod "d3de536b-47a0-4884-b6fb-3021b4ecbb98" (UID: "d3de536b-47a0-4884-b6fb-3021b4ecbb98"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.693656 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaf83795-11a7-477a-9ecb-a2659ce4deda-config-data" (OuterVolumeSpecName: "config-data") pod "eaf83795-11a7-477a-9ecb-a2659ce4deda" (UID: "eaf83795-11a7-477a-9ecb-a2659ce4deda"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.705859 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d3de536b-47a0-4884-b6fb-3021b4ecbb98" (UID: "d3de536b-47a0-4884-b6fb-3021b4ecbb98"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.741827 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaf83795-11a7-477a-9ecb-a2659ce4deda-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eaf83795-11a7-477a-9ecb-a2659ce4deda" (UID: "eaf83795-11a7-477a-9ecb-a2659ce4deda"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.757332 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3de536b-47a0-4884-b6fb-3021b4ecbb98" (UID: "d3de536b-47a0-4884-b6fb-3021b4ecbb98"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.761714 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf83795-11a7-477a-9ecb-a2659ce4deda-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.761818 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lct6l\" (UniqueName: \"kubernetes.io/projected/eaf83795-11a7-477a-9ecb-a2659ce4deda-kube-api-access-lct6l\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.761834 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.761845 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.761855 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89gfv\" (UniqueName: \"kubernetes.io/projected/d3de536b-47a0-4884-b6fb-3021b4ecbb98-kube-api-access-89gfv\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.761866 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eaf83795-11a7-477a-9ecb-a2659ce4deda-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.761877 4884 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.761888 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaf83795-11a7-477a-9ecb-a2659ce4deda-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.773332 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d3de536b-47a0-4884-b6fb-3021b4ecbb98" (UID: "d3de536b-47a0-4884-b6fb-3021b4ecbb98"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.806896 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-config-data" (OuterVolumeSpecName: "config-data") pod "d3de536b-47a0-4884-b6fb-3021b4ecbb98" (UID: "d3de536b-47a0-4884-b6fb-3021b4ecbb98"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.863574 4884 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.863607 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3de536b-47a0-4884-b6fb-3021b4ecbb98-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.884535 4884 generic.go:334] "Generic (PLEG): container finished" podID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerID="d83ec64874b2bf694026e9c2b12dd39ab9de100fdb6207cde5e15afad02cab9f" exitCode=0 Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.884612 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.884607 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3de536b-47a0-4884-b6fb-3021b4ecbb98","Type":"ContainerDied","Data":"d83ec64874b2bf694026e9c2b12dd39ab9de100fdb6207cde5e15afad02cab9f"} Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.884699 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d3de536b-47a0-4884-b6fb-3021b4ecbb98","Type":"ContainerDied","Data":"2b278c38f70621e92e55bb2dea375dbf45c48c9a7fec8593d85c6672cc4f7ca6"} Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.884737 4884 scope.go:117] "RemoveContainer" containerID="d7821da0426d4142c952cc46bc599f50b094801e95a7e2c96baa1a1f215e33cf" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.886477 4884 generic.go:334] "Generic (PLEG): container finished" podID="eaf83795-11a7-477a-9ecb-a2659ce4deda" containerID="a334be3738065008f1cc04000ba8e76c1c0cf8fc7b89df9cc376d227b0071fff" exitCode=0 Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.886513 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.886525 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eaf83795-11a7-477a-9ecb-a2659ce4deda","Type":"ContainerDied","Data":"a334be3738065008f1cc04000ba8e76c1c0cf8fc7b89df9cc376d227b0071fff"} Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.886551 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"eaf83795-11a7-477a-9ecb-a2659ce4deda","Type":"ContainerDied","Data":"1de553a8de08fd0a8f3b7c4b482318858805412993c85abf07293413a9613402"} Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.934007 4884 scope.go:117] "RemoveContainer" containerID="56dc8361d659338a5ec8ff240b527700ca795747e611ad29313581993b3fe876" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.944489 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.954589 4884 scope.go:117] "RemoveContainer" containerID="d83ec64874b2bf694026e9c2b12dd39ab9de100fdb6207cde5e15afad02cab9f" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.966451 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.983916 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.986440 4884 scope.go:117] "RemoveContainer" containerID="59c87f346db8a4066aad1039e1b53d620c6bac171e09b11b40c5e62095f3c2d0" Dec 02 01:59:16 crc kubenswrapper[4884]: I1202 01:59:16.993902 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.005466 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 02 01:59:17 crc kubenswrapper[4884]: E1202 01:59:17.005904 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf83795-11a7-477a-9ecb-a2659ce4deda" containerName="nova-api-log" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.005916 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf83795-11a7-477a-9ecb-a2659ce4deda" containerName="nova-api-log" Dec 02 01:59:17 crc kubenswrapper[4884]: E1202 01:59:17.005925 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="ceilometer-notification-agent" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.005931 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="ceilometer-notification-agent" Dec 02 01:59:17 crc kubenswrapper[4884]: E1202 01:59:17.005946 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="ceilometer-central-agent" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.005952 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="ceilometer-central-agent" Dec 02 01:59:17 crc kubenswrapper[4884]: E1202 01:59:17.005961 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf83795-11a7-477a-9ecb-a2659ce4deda" containerName="nova-api-api" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.005966 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf83795-11a7-477a-9ecb-a2659ce4deda" containerName="nova-api-api" Dec 02 01:59:17 crc kubenswrapper[4884]: E1202 01:59:17.005984 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="proxy-httpd" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.005989 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="proxy-httpd" Dec 02 01:59:17 crc kubenswrapper[4884]: E1202 01:59:17.006006 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="sg-core" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.006012 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="sg-core" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.006206 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="ceilometer-notification-agent" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.006218 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="ceilometer-central-agent" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.006242 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf83795-11a7-477a-9ecb-a2659ce4deda" containerName="nova-api-api" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.006254 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="proxy-httpd" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.006269 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf83795-11a7-477a-9ecb-a2659ce4deda" containerName="nova-api-log" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.006285 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" containerName="sg-core" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.007314 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.011485 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.011763 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.011840 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.016802 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.019440 4884 scope.go:117] "RemoveContainer" containerID="d7821da0426d4142c952cc46bc599f50b094801e95a7e2c96baa1a1f215e33cf" Dec 02 01:59:17 crc kubenswrapper[4884]: E1202 01:59:17.020241 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7821da0426d4142c952cc46bc599f50b094801e95a7e2c96baa1a1f215e33cf\": container with ID starting with d7821da0426d4142c952cc46bc599f50b094801e95a7e2c96baa1a1f215e33cf not found: ID does not exist" containerID="d7821da0426d4142c952cc46bc599f50b094801e95a7e2c96baa1a1f215e33cf" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.020270 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7821da0426d4142c952cc46bc599f50b094801e95a7e2c96baa1a1f215e33cf"} err="failed to get container status \"d7821da0426d4142c952cc46bc599f50b094801e95a7e2c96baa1a1f215e33cf\": rpc error: code = NotFound desc = could not find container \"d7821da0426d4142c952cc46bc599f50b094801e95a7e2c96baa1a1f215e33cf\": container with ID starting with d7821da0426d4142c952cc46bc599f50b094801e95a7e2c96baa1a1f215e33cf not found: ID does not exist" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.020298 4884 scope.go:117] "RemoveContainer" containerID="56dc8361d659338a5ec8ff240b527700ca795747e611ad29313581993b3fe876" Dec 02 01:59:17 crc kubenswrapper[4884]: E1202 01:59:17.020526 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56dc8361d659338a5ec8ff240b527700ca795747e611ad29313581993b3fe876\": container with ID starting with 56dc8361d659338a5ec8ff240b527700ca795747e611ad29313581993b3fe876 not found: ID does not exist" containerID="56dc8361d659338a5ec8ff240b527700ca795747e611ad29313581993b3fe876" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.020556 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56dc8361d659338a5ec8ff240b527700ca795747e611ad29313581993b3fe876"} err="failed to get container status \"56dc8361d659338a5ec8ff240b527700ca795747e611ad29313581993b3fe876\": rpc error: code = NotFound desc = could not find container \"56dc8361d659338a5ec8ff240b527700ca795747e611ad29313581993b3fe876\": container with ID starting with 56dc8361d659338a5ec8ff240b527700ca795747e611ad29313581993b3fe876 not found: ID does not exist" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.020576 4884 scope.go:117] "RemoveContainer" containerID="d83ec64874b2bf694026e9c2b12dd39ab9de100fdb6207cde5e15afad02cab9f" Dec 02 01:59:17 crc kubenswrapper[4884]: E1202 01:59:17.020836 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d83ec64874b2bf694026e9c2b12dd39ab9de100fdb6207cde5e15afad02cab9f\": container with ID starting with d83ec64874b2bf694026e9c2b12dd39ab9de100fdb6207cde5e15afad02cab9f not found: ID does not exist" containerID="d83ec64874b2bf694026e9c2b12dd39ab9de100fdb6207cde5e15afad02cab9f" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.020863 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d83ec64874b2bf694026e9c2b12dd39ab9de100fdb6207cde5e15afad02cab9f"} err="failed to get container status \"d83ec64874b2bf694026e9c2b12dd39ab9de100fdb6207cde5e15afad02cab9f\": rpc error: code = NotFound desc = could not find container \"d83ec64874b2bf694026e9c2b12dd39ab9de100fdb6207cde5e15afad02cab9f\": container with ID starting with d83ec64874b2bf694026e9c2b12dd39ab9de100fdb6207cde5e15afad02cab9f not found: ID does not exist" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.020890 4884 scope.go:117] "RemoveContainer" containerID="59c87f346db8a4066aad1039e1b53d620c6bac171e09b11b40c5e62095f3c2d0" Dec 02 01:59:17 crc kubenswrapper[4884]: E1202 01:59:17.026246 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59c87f346db8a4066aad1039e1b53d620c6bac171e09b11b40c5e62095f3c2d0\": container with ID starting with 59c87f346db8a4066aad1039e1b53d620c6bac171e09b11b40c5e62095f3c2d0 not found: ID does not exist" containerID="59c87f346db8a4066aad1039e1b53d620c6bac171e09b11b40c5e62095f3c2d0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.026293 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59c87f346db8a4066aad1039e1b53d620c6bac171e09b11b40c5e62095f3c2d0"} err="failed to get container status \"59c87f346db8a4066aad1039e1b53d620c6bac171e09b11b40c5e62095f3c2d0\": rpc error: code = NotFound desc = could not find container \"59c87f346db8a4066aad1039e1b53d620c6bac171e09b11b40c5e62095f3c2d0\": container with ID starting with 59c87f346db8a4066aad1039e1b53d620c6bac171e09b11b40c5e62095f3c2d0 not found: ID does not exist" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.026324 4884 scope.go:117] "RemoveContainer" containerID="a334be3738065008f1cc04000ba8e76c1c0cf8fc7b89df9cc376d227b0071fff" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.031580 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.035048 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.037714 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.037784 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.038889 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.046504 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.050147 4884 scope.go:117] "RemoveContainer" containerID="be1699158e9ea560fdd25573e7c4baab0c870736e5eb55256359d82f90050ef6" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.066611 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.066651 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.066674 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.066807 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-scripts\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.066835 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-log-httpd\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.066873 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-internal-tls-certs\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.066969 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-config-data\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.067028 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nxsz\" (UniqueName: \"kubernetes.io/projected/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-kube-api-access-8nxsz\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.067051 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmdgp\" (UniqueName: \"kubernetes.io/projected/194a7ea1-a1f9-4868-9acc-628f909b9b11-kube-api-access-jmdgp\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.067075 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-config-data\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.067149 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-run-httpd\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.067182 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/194a7ea1-a1f9-4868-9acc-628f909b9b11-logs\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.067238 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.067369 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-public-tls-certs\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.074018 4884 scope.go:117] "RemoveContainer" containerID="a334be3738065008f1cc04000ba8e76c1c0cf8fc7b89df9cc376d227b0071fff" Dec 02 01:59:17 crc kubenswrapper[4884]: E1202 01:59:17.074459 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a334be3738065008f1cc04000ba8e76c1c0cf8fc7b89df9cc376d227b0071fff\": container with ID starting with a334be3738065008f1cc04000ba8e76c1c0cf8fc7b89df9cc376d227b0071fff not found: ID does not exist" containerID="a334be3738065008f1cc04000ba8e76c1c0cf8fc7b89df9cc376d227b0071fff" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.074499 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a334be3738065008f1cc04000ba8e76c1c0cf8fc7b89df9cc376d227b0071fff"} err="failed to get container status \"a334be3738065008f1cc04000ba8e76c1c0cf8fc7b89df9cc376d227b0071fff\": rpc error: code = NotFound desc = could not find container \"a334be3738065008f1cc04000ba8e76c1c0cf8fc7b89df9cc376d227b0071fff\": container with ID starting with a334be3738065008f1cc04000ba8e76c1c0cf8fc7b89df9cc376d227b0071fff not found: ID does not exist" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.074526 4884 scope.go:117] "RemoveContainer" containerID="be1699158e9ea560fdd25573e7c4baab0c870736e5eb55256359d82f90050ef6" Dec 02 01:59:17 crc kubenswrapper[4884]: E1202 01:59:17.074860 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be1699158e9ea560fdd25573e7c4baab0c870736e5eb55256359d82f90050ef6\": container with ID starting with be1699158e9ea560fdd25573e7c4baab0c870736e5eb55256359d82f90050ef6 not found: ID does not exist" containerID="be1699158e9ea560fdd25573e7c4baab0c870736e5eb55256359d82f90050ef6" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.074895 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be1699158e9ea560fdd25573e7c4baab0c870736e5eb55256359d82f90050ef6"} err="failed to get container status \"be1699158e9ea560fdd25573e7c4baab0c870736e5eb55256359d82f90050ef6\": rpc error: code = NotFound desc = could not find container \"be1699158e9ea560fdd25573e7c4baab0c870736e5eb55256359d82f90050ef6\": container with ID starting with be1699158e9ea560fdd25573e7c4baab0c870736e5eb55256359d82f90050ef6 not found: ID does not exist" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.169969 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-internal-tls-certs\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.170104 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-config-data\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.170197 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nxsz\" (UniqueName: \"kubernetes.io/projected/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-kube-api-access-8nxsz\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.170263 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmdgp\" (UniqueName: \"kubernetes.io/projected/194a7ea1-a1f9-4868-9acc-628f909b9b11-kube-api-access-jmdgp\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.170287 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-config-data\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.170378 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-run-httpd\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.170443 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/194a7ea1-a1f9-4868-9acc-628f909b9b11-logs\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.170506 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.170584 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-public-tls-certs\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.170616 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.170660 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.170686 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.170871 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-scripts\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.170946 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-log-httpd\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.171021 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/194a7ea1-a1f9-4868-9acc-628f909b9b11-logs\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.171065 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-run-httpd\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.171608 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-log-httpd\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.174206 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.175229 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-scripts\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.175459 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.175515 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.176245 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-internal-tls-certs\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.177678 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-config-data\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.179283 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-config-data\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.179985 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.180496 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-public-tls-certs\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.188814 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmdgp\" (UniqueName: \"kubernetes.io/projected/194a7ea1-a1f9-4868-9acc-628f909b9b11-kube-api-access-jmdgp\") pod \"nova-api-0\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.192438 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nxsz\" (UniqueName: \"kubernetes.io/projected/0bfa748c-a54a-4a8c-bae8-f87fefe8a81a-kube-api-access-8nxsz\") pod \"ceilometer-0\" (UID: \"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a\") " pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.330170 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.351613 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.626475 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3de536b-47a0-4884-b6fb-3021b4ecbb98" path="/var/lib/kubelet/pods/d3de536b-47a0-4884-b6fb-3021b4ecbb98/volumes" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.627523 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaf83795-11a7-477a-9ecb-a2659ce4deda" path="/var/lib/kubelet/pods/eaf83795-11a7-477a-9ecb-a2659ce4deda/volumes" Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.885559 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.905925 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 02 01:59:17 crc kubenswrapper[4884]: I1202 01:59:17.919547 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"194a7ea1-a1f9-4868-9acc-628f909b9b11","Type":"ContainerStarted","Data":"e0e6ed378b30a307833cc137fb11a42135844e10557df57ee98e796c6ff5c6a2"} Dec 02 01:59:17 crc kubenswrapper[4884]: W1202 01:59:17.926512 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0bfa748c_a54a_4a8c_bae8_f87fefe8a81a.slice/crio-62d682d9cb194753fd3b1b3886e34f27d53cc3e96e9dbb5d69ed2d0edc273cb3 WatchSource:0}: Error finding container 62d682d9cb194753fd3b1b3886e34f27d53cc3e96e9dbb5d69ed2d0edc273cb3: Status 404 returned error can't find the container with id 62d682d9cb194753fd3b1b3886e34f27d53cc3e96e9dbb5d69ed2d0edc273cb3 Dec 02 01:59:18 crc kubenswrapper[4884]: I1202 01:59:18.942371 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"194a7ea1-a1f9-4868-9acc-628f909b9b11","Type":"ContainerStarted","Data":"67e389f9a1bbeff96e6d5b65b6cf85198346fa8edfe1cce938dc8a555607f6f1"} Dec 02 01:59:18 crc kubenswrapper[4884]: I1202 01:59:18.942728 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"194a7ea1-a1f9-4868-9acc-628f909b9b11","Type":"ContainerStarted","Data":"b8caabff5dc73ebcb399f862b4f0319e13bc01e5b8b668c981308df68c05fd5d"} Dec 02 01:59:18 crc kubenswrapper[4884]: I1202 01:59:18.951543 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a","Type":"ContainerStarted","Data":"91415173dab3bf1802fc85df489da64f699b5b03dd90dcc16c451095ecde6650"} Dec 02 01:59:18 crc kubenswrapper[4884]: I1202 01:59:18.951593 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a","Type":"ContainerStarted","Data":"62d682d9cb194753fd3b1b3886e34f27d53cc3e96e9dbb5d69ed2d0edc273cb3"} Dec 02 01:59:18 crc kubenswrapper[4884]: I1202 01:59:18.972852 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.972827182 podStartE2EDuration="2.972827182s" podCreationTimestamp="2025-12-02 01:59:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:59:18.964876606 +0000 UTC m=+1375.640713530" watchObservedRunningTime="2025-12-02 01:59:18.972827182 +0000 UTC m=+1375.648664076" Dec 02 01:59:19 crc kubenswrapper[4884]: I1202 01:59:19.963379 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a","Type":"ContainerStarted","Data":"de4675f2d1d3f3e77828af332f4463de238478509cfc53bc229bb8e5f30bf08c"} Dec 02 01:59:20 crc kubenswrapper[4884]: I1202 01:59:20.183692 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:20 crc kubenswrapper[4884]: I1202 01:59:20.210686 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:20 crc kubenswrapper[4884]: I1202 01:59:20.791714 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 01:59:20 crc kubenswrapper[4884]: I1202 01:59:20.875151 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-b74v6"] Dec 02 01:59:20 crc kubenswrapper[4884]: I1202 01:59:20.879068 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-b74v6" podUID="2686aabc-d1de-4d56-95db-0b919d055b3d" containerName="dnsmasq-dns" containerID="cri-o://24be14cd74684b8efe78f4efdb6cc002f96bdfb7ffe408549db0620403f46436" gracePeriod=10 Dec 02 01:59:20 crc kubenswrapper[4884]: I1202 01:59:20.982372 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a","Type":"ContainerStarted","Data":"986a808244a4d746a1681f9e581ed8e56dc54154af101be47af2569550e5bc56"} Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.006381 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.214392 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-djwlt"] Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.215693 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.218091 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.218509 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.227927 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-djwlt"] Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.253702 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n56h\" (UniqueName: \"kubernetes.io/projected/1acb68f2-bed9-462f-82eb-244f84069402-kube-api-access-5n56h\") pod \"nova-cell1-cell-mapping-djwlt\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.253810 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-config-data\") pod \"nova-cell1-cell-mapping-djwlt\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.253841 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-scripts\") pod \"nova-cell1-cell-mapping-djwlt\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.253916 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-djwlt\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.366557 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-djwlt\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.366663 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n56h\" (UniqueName: \"kubernetes.io/projected/1acb68f2-bed9-462f-82eb-244f84069402-kube-api-access-5n56h\") pod \"nova-cell1-cell-mapping-djwlt\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.366766 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-config-data\") pod \"nova-cell1-cell-mapping-djwlt\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.366813 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-scripts\") pod \"nova-cell1-cell-mapping-djwlt\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.393779 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-config-data\") pod \"nova-cell1-cell-mapping-djwlt\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.399426 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-scripts\") pod \"nova-cell1-cell-mapping-djwlt\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.408016 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-djwlt\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.410576 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n56h\" (UniqueName: \"kubernetes.io/projected/1acb68f2-bed9-462f-82eb-244f84069402-kube-api-access-5n56h\") pod \"nova-cell1-cell-mapping-djwlt\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.465902 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.535075 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.572619 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-ovsdbserver-sb\") pod \"2686aabc-d1de-4d56-95db-0b919d055b3d\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.572683 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-ovsdbserver-nb\") pod \"2686aabc-d1de-4d56-95db-0b919d055b3d\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.572782 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sw9bv\" (UniqueName: \"kubernetes.io/projected/2686aabc-d1de-4d56-95db-0b919d055b3d-kube-api-access-sw9bv\") pod \"2686aabc-d1de-4d56-95db-0b919d055b3d\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.572878 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-dns-svc\") pod \"2686aabc-d1de-4d56-95db-0b919d055b3d\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.572910 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-dns-swift-storage-0\") pod \"2686aabc-d1de-4d56-95db-0b919d055b3d\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.572969 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-config\") pod \"2686aabc-d1de-4d56-95db-0b919d055b3d\" (UID: \"2686aabc-d1de-4d56-95db-0b919d055b3d\") " Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.581538 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2686aabc-d1de-4d56-95db-0b919d055b3d-kube-api-access-sw9bv" (OuterVolumeSpecName: "kube-api-access-sw9bv") pod "2686aabc-d1de-4d56-95db-0b919d055b3d" (UID: "2686aabc-d1de-4d56-95db-0b919d055b3d"). InnerVolumeSpecName "kube-api-access-sw9bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.671433 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2686aabc-d1de-4d56-95db-0b919d055b3d" (UID: "2686aabc-d1de-4d56-95db-0b919d055b3d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.676476 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.676512 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sw9bv\" (UniqueName: \"kubernetes.io/projected/2686aabc-d1de-4d56-95db-0b919d055b3d-kube-api-access-sw9bv\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.682233 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2686aabc-d1de-4d56-95db-0b919d055b3d" (UID: "2686aabc-d1de-4d56-95db-0b919d055b3d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.682404 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-config" (OuterVolumeSpecName: "config") pod "2686aabc-d1de-4d56-95db-0b919d055b3d" (UID: "2686aabc-d1de-4d56-95db-0b919d055b3d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.687944 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2686aabc-d1de-4d56-95db-0b919d055b3d" (UID: "2686aabc-d1de-4d56-95db-0b919d055b3d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.729794 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2686aabc-d1de-4d56-95db-0b919d055b3d" (UID: "2686aabc-d1de-4d56-95db-0b919d055b3d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.777063 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.777091 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.777101 4884 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.777109 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2686aabc-d1de-4d56-95db-0b919d055b3d-config\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.994078 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0bfa748c-a54a-4a8c-bae8-f87fefe8a81a","Type":"ContainerStarted","Data":"117907da2d178749abaf63886b0414ca1efc9174f6d23b18aa50b1f422fe7a90"} Dec 02 01:59:21 crc kubenswrapper[4884]: I1202 01:59:21.994224 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 02 01:59:22 crc kubenswrapper[4884]: I1202 01:59:21.996686 4884 generic.go:334] "Generic (PLEG): container finished" podID="2686aabc-d1de-4d56-95db-0b919d055b3d" containerID="24be14cd74684b8efe78f4efdb6cc002f96bdfb7ffe408549db0620403f46436" exitCode=0 Dec 02 01:59:22 crc kubenswrapper[4884]: I1202 01:59:21.996770 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-b74v6" Dec 02 01:59:22 crc kubenswrapper[4884]: I1202 01:59:21.996807 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-b74v6" event={"ID":"2686aabc-d1de-4d56-95db-0b919d055b3d","Type":"ContainerDied","Data":"24be14cd74684b8efe78f4efdb6cc002f96bdfb7ffe408549db0620403f46436"} Dec 02 01:59:22 crc kubenswrapper[4884]: I1202 01:59:21.996847 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-b74v6" event={"ID":"2686aabc-d1de-4d56-95db-0b919d055b3d","Type":"ContainerDied","Data":"6b5585bce3361f67293c467b46b6db49399765fadfcc4deefd1a020991104bcf"} Dec 02 01:59:22 crc kubenswrapper[4884]: I1202 01:59:21.996871 4884 scope.go:117] "RemoveContainer" containerID="24be14cd74684b8efe78f4efdb6cc002f96bdfb7ffe408549db0620403f46436" Dec 02 01:59:22 crc kubenswrapper[4884]: I1202 01:59:22.030763 4884 scope.go:117] "RemoveContainer" containerID="e70c728f8da1585c72cc4e4a885c9377b7f2957a5aac621406e67efdea96a6b6" Dec 02 01:59:22 crc kubenswrapper[4884]: I1202 01:59:22.046390 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.410170813 podStartE2EDuration="6.04636647s" podCreationTimestamp="2025-12-02 01:59:16 +0000 UTC" firstStartedPulling="2025-12-02 01:59:17.928500719 +0000 UTC m=+1374.604337613" lastFinishedPulling="2025-12-02 01:59:21.564696386 +0000 UTC m=+1378.240533270" observedRunningTime="2025-12-02 01:59:22.030261253 +0000 UTC m=+1378.706098137" watchObservedRunningTime="2025-12-02 01:59:22.04636647 +0000 UTC m=+1378.722203364" Dec 02 01:59:22 crc kubenswrapper[4884]: I1202 01:59:22.060781 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-b74v6"] Dec 02 01:59:22 crc kubenswrapper[4884]: I1202 01:59:22.082360 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-djwlt"] Dec 02 01:59:22 crc kubenswrapper[4884]: I1202 01:59:22.098143 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-b74v6"] Dec 02 01:59:22 crc kubenswrapper[4884]: I1202 01:59:22.101342 4884 scope.go:117] "RemoveContainer" containerID="24be14cd74684b8efe78f4efdb6cc002f96bdfb7ffe408549db0620403f46436" Dec 02 01:59:22 crc kubenswrapper[4884]: E1202 01:59:22.101688 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24be14cd74684b8efe78f4efdb6cc002f96bdfb7ffe408549db0620403f46436\": container with ID starting with 24be14cd74684b8efe78f4efdb6cc002f96bdfb7ffe408549db0620403f46436 not found: ID does not exist" containerID="24be14cd74684b8efe78f4efdb6cc002f96bdfb7ffe408549db0620403f46436" Dec 02 01:59:22 crc kubenswrapper[4884]: I1202 01:59:22.101717 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24be14cd74684b8efe78f4efdb6cc002f96bdfb7ffe408549db0620403f46436"} err="failed to get container status \"24be14cd74684b8efe78f4efdb6cc002f96bdfb7ffe408549db0620403f46436\": rpc error: code = NotFound desc = could not find container \"24be14cd74684b8efe78f4efdb6cc002f96bdfb7ffe408549db0620403f46436\": container with ID starting with 24be14cd74684b8efe78f4efdb6cc002f96bdfb7ffe408549db0620403f46436 not found: ID does not exist" Dec 02 01:59:22 crc kubenswrapper[4884]: I1202 01:59:22.101736 4884 scope.go:117] "RemoveContainer" containerID="e70c728f8da1585c72cc4e4a885c9377b7f2957a5aac621406e67efdea96a6b6" Dec 02 01:59:22 crc kubenswrapper[4884]: E1202 01:59:22.102043 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e70c728f8da1585c72cc4e4a885c9377b7f2957a5aac621406e67efdea96a6b6\": container with ID starting with e70c728f8da1585c72cc4e4a885c9377b7f2957a5aac621406e67efdea96a6b6 not found: ID does not exist" containerID="e70c728f8da1585c72cc4e4a885c9377b7f2957a5aac621406e67efdea96a6b6" Dec 02 01:59:22 crc kubenswrapper[4884]: I1202 01:59:22.102065 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e70c728f8da1585c72cc4e4a885c9377b7f2957a5aac621406e67efdea96a6b6"} err="failed to get container status \"e70c728f8da1585c72cc4e4a885c9377b7f2957a5aac621406e67efdea96a6b6\": rpc error: code = NotFound desc = could not find container \"e70c728f8da1585c72cc4e4a885c9377b7f2957a5aac621406e67efdea96a6b6\": container with ID starting with e70c728f8da1585c72cc4e4a885c9377b7f2957a5aac621406e67efdea96a6b6 not found: ID does not exist" Dec 02 01:59:23 crc kubenswrapper[4884]: I1202 01:59:23.006077 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-djwlt" event={"ID":"1acb68f2-bed9-462f-82eb-244f84069402","Type":"ContainerStarted","Data":"ab1060d0fe9f461502ed0a14aecefbdef801e68eae01d8e5e79d6aa38891af6d"} Dec 02 01:59:23 crc kubenswrapper[4884]: I1202 01:59:23.006559 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-djwlt" event={"ID":"1acb68f2-bed9-462f-82eb-244f84069402","Type":"ContainerStarted","Data":"b702874cd67ad41b0a53f0a197e647dc77bfd0039300326f7ca97f3a4bd0bf8e"} Dec 02 01:59:23 crc kubenswrapper[4884]: I1202 01:59:23.030032 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-djwlt" podStartSLOduration=2.030016259 podStartE2EDuration="2.030016259s" podCreationTimestamp="2025-12-02 01:59:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:59:23.024196216 +0000 UTC m=+1379.700033120" watchObservedRunningTime="2025-12-02 01:59:23.030016259 +0000 UTC m=+1379.705853143" Dec 02 01:59:23 crc kubenswrapper[4884]: I1202 01:59:23.633050 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2686aabc-d1de-4d56-95db-0b919d055b3d" path="/var/lib/kubelet/pods/2686aabc-d1de-4d56-95db-0b919d055b3d/volumes" Dec 02 01:59:27 crc kubenswrapper[4884]: I1202 01:59:27.057789 4884 generic.go:334] "Generic (PLEG): container finished" podID="1acb68f2-bed9-462f-82eb-244f84069402" containerID="ab1060d0fe9f461502ed0a14aecefbdef801e68eae01d8e5e79d6aa38891af6d" exitCode=0 Dec 02 01:59:27 crc kubenswrapper[4884]: I1202 01:59:27.057891 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-djwlt" event={"ID":"1acb68f2-bed9-462f-82eb-244f84069402","Type":"ContainerDied","Data":"ab1060d0fe9f461502ed0a14aecefbdef801e68eae01d8e5e79d6aa38891af6d"} Dec 02 01:59:27 crc kubenswrapper[4884]: I1202 01:59:27.330858 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 01:59:27 crc kubenswrapper[4884]: I1202 01:59:27.330941 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 01:59:28 crc kubenswrapper[4884]: I1202 01:59:28.347981 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="194a7ea1-a1f9-4868-9acc-628f909b9b11" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.216:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 01:59:28 crc kubenswrapper[4884]: I1202 01:59:28.347996 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="194a7ea1-a1f9-4868-9acc-628f909b9b11" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.216:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 01:59:28 crc kubenswrapper[4884]: I1202 01:59:28.410154 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:28 crc kubenswrapper[4884]: I1202 01:59:28.467839 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n56h\" (UniqueName: \"kubernetes.io/projected/1acb68f2-bed9-462f-82eb-244f84069402-kube-api-access-5n56h\") pod \"1acb68f2-bed9-462f-82eb-244f84069402\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " Dec 02 01:59:28 crc kubenswrapper[4884]: I1202 01:59:28.471626 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-scripts\") pod \"1acb68f2-bed9-462f-82eb-244f84069402\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " Dec 02 01:59:28 crc kubenswrapper[4884]: I1202 01:59:28.472179 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-combined-ca-bundle\") pod \"1acb68f2-bed9-462f-82eb-244f84069402\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " Dec 02 01:59:28 crc kubenswrapper[4884]: I1202 01:59:28.472349 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-config-data\") pod \"1acb68f2-bed9-462f-82eb-244f84069402\" (UID: \"1acb68f2-bed9-462f-82eb-244f84069402\") " Dec 02 01:59:28 crc kubenswrapper[4884]: I1202 01:59:28.486952 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-scripts" (OuterVolumeSpecName: "scripts") pod "1acb68f2-bed9-462f-82eb-244f84069402" (UID: "1acb68f2-bed9-462f-82eb-244f84069402"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:28 crc kubenswrapper[4884]: I1202 01:59:28.493031 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1acb68f2-bed9-462f-82eb-244f84069402-kube-api-access-5n56h" (OuterVolumeSpecName: "kube-api-access-5n56h") pod "1acb68f2-bed9-462f-82eb-244f84069402" (UID: "1acb68f2-bed9-462f-82eb-244f84069402"). InnerVolumeSpecName "kube-api-access-5n56h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:59:28 crc kubenswrapper[4884]: I1202 01:59:28.520791 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1acb68f2-bed9-462f-82eb-244f84069402" (UID: "1acb68f2-bed9-462f-82eb-244f84069402"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:28 crc kubenswrapper[4884]: I1202 01:59:28.530840 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-config-data" (OuterVolumeSpecName: "config-data") pod "1acb68f2-bed9-462f-82eb-244f84069402" (UID: "1acb68f2-bed9-462f-82eb-244f84069402"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:28 crc kubenswrapper[4884]: I1202 01:59:28.576083 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n56h\" (UniqueName: \"kubernetes.io/projected/1acb68f2-bed9-462f-82eb-244f84069402-kube-api-access-5n56h\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:28 crc kubenswrapper[4884]: I1202 01:59:28.576111 4884 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-scripts\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:28 crc kubenswrapper[4884]: I1202 01:59:28.576121 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:28 crc kubenswrapper[4884]: I1202 01:59:28.576130 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1acb68f2-bed9-462f-82eb-244f84069402-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:29 crc kubenswrapper[4884]: I1202 01:59:29.084652 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-djwlt" event={"ID":"1acb68f2-bed9-462f-82eb-244f84069402","Type":"ContainerDied","Data":"b702874cd67ad41b0a53f0a197e647dc77bfd0039300326f7ca97f3a4bd0bf8e"} Dec 02 01:59:29 crc kubenswrapper[4884]: I1202 01:59:29.085317 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b702874cd67ad41b0a53f0a197e647dc77bfd0039300326f7ca97f3a4bd0bf8e" Dec 02 01:59:29 crc kubenswrapper[4884]: I1202 01:59:29.085340 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-djwlt" Dec 02 01:59:29 crc kubenswrapper[4884]: I1202 01:59:29.326771 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:59:29 crc kubenswrapper[4884]: I1202 01:59:29.327629 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="194a7ea1-a1f9-4868-9acc-628f909b9b11" containerName="nova-api-api" containerID="cri-o://67e389f9a1bbeff96e6d5b65b6cf85198346fa8edfe1cce938dc8a555607f6f1" gracePeriod=30 Dec 02 01:59:29 crc kubenswrapper[4884]: I1202 01:59:29.328039 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="194a7ea1-a1f9-4868-9acc-628f909b9b11" containerName="nova-api-log" containerID="cri-o://b8caabff5dc73ebcb399f862b4f0319e13bc01e5b8b668c981308df68c05fd5d" gracePeriod=30 Dec 02 01:59:29 crc kubenswrapper[4884]: I1202 01:59:29.351615 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 01:59:29 crc kubenswrapper[4884]: I1202 01:59:29.351859 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d17a0f0d-1e28-4238-83f8-742b3b3df02c" containerName="nova-scheduler-scheduler" containerID="cri-o://6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd" gracePeriod=30 Dec 02 01:59:29 crc kubenswrapper[4884]: I1202 01:59:29.398387 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:59:29 crc kubenswrapper[4884]: I1202 01:59:29.398623 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="062eea62-0f0f-45c4-a2e2-24820062d522" containerName="nova-metadata-log" containerID="cri-o://5eba1d389242792d3f6698b0d5b8aa88025a5f1c27ab83416aa209c596ce5ba5" gracePeriod=30 Dec 02 01:59:29 crc kubenswrapper[4884]: I1202 01:59:29.398996 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="062eea62-0f0f-45c4-a2e2-24820062d522" containerName="nova-metadata-metadata" containerID="cri-o://dcbd98665831ff5ed42a0debbf06d1ce2e3efeff9859ff024feeae70f3ac0952" gracePeriod=30 Dec 02 01:59:30 crc kubenswrapper[4884]: I1202 01:59:30.095501 4884 generic.go:334] "Generic (PLEG): container finished" podID="194a7ea1-a1f9-4868-9acc-628f909b9b11" containerID="b8caabff5dc73ebcb399f862b4f0319e13bc01e5b8b668c981308df68c05fd5d" exitCode=143 Dec 02 01:59:30 crc kubenswrapper[4884]: I1202 01:59:30.095565 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"194a7ea1-a1f9-4868-9acc-628f909b9b11","Type":"ContainerDied","Data":"b8caabff5dc73ebcb399f862b4f0319e13bc01e5b8b668c981308df68c05fd5d"} Dec 02 01:59:30 crc kubenswrapper[4884]: I1202 01:59:30.097777 4884 generic.go:334] "Generic (PLEG): container finished" podID="062eea62-0f0f-45c4-a2e2-24820062d522" containerID="5eba1d389242792d3f6698b0d5b8aa88025a5f1c27ab83416aa209c596ce5ba5" exitCode=143 Dec 02 01:59:30 crc kubenswrapper[4884]: I1202 01:59:30.097801 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"062eea62-0f0f-45c4-a2e2-24820062d522","Type":"ContainerDied","Data":"5eba1d389242792d3f6698b0d5b8aa88025a5f1c27ab83416aa209c596ce5ba5"} Dec 02 01:59:30 crc kubenswrapper[4884]: E1202 01:59:30.910146 4884 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 02 01:59:30 crc kubenswrapper[4884]: E1202 01:59:30.912326 4884 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 02 01:59:30 crc kubenswrapper[4884]: E1202 01:59:30.914406 4884 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 02 01:59:30 crc kubenswrapper[4884]: E1202 01:59:30.914442 4884 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="d17a0f0d-1e28-4238-83f8-742b3b3df02c" containerName="nova-scheduler-scheduler" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.069010 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.167352 4884 generic.go:334] "Generic (PLEG): container finished" podID="062eea62-0f0f-45c4-a2e2-24820062d522" containerID="dcbd98665831ff5ed42a0debbf06d1ce2e3efeff9859ff024feeae70f3ac0952" exitCode=0 Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.167619 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"062eea62-0f0f-45c4-a2e2-24820062d522","Type":"ContainerDied","Data":"dcbd98665831ff5ed42a0debbf06d1ce2e3efeff9859ff024feeae70f3ac0952"} Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.167644 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"062eea62-0f0f-45c4-a2e2-24820062d522","Type":"ContainerDied","Data":"50e67ea73a7cfa5e4e71d3a9b63e955178eac72b584d3c64dd0155facae7532c"} Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.167671 4884 scope.go:117] "RemoveContainer" containerID="dcbd98665831ff5ed42a0debbf06d1ce2e3efeff9859ff024feeae70f3ac0952" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.167811 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.175965 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-nova-metadata-tls-certs\") pod \"062eea62-0f0f-45c4-a2e2-24820062d522\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.176012 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-combined-ca-bundle\") pod \"062eea62-0f0f-45c4-a2e2-24820062d522\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.176127 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/062eea62-0f0f-45c4-a2e2-24820062d522-logs\") pod \"062eea62-0f0f-45c4-a2e2-24820062d522\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.176171 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-config-data\") pod \"062eea62-0f0f-45c4-a2e2-24820062d522\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.176189 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vg4x9\" (UniqueName: \"kubernetes.io/projected/062eea62-0f0f-45c4-a2e2-24820062d522-kube-api-access-vg4x9\") pod \"062eea62-0f0f-45c4-a2e2-24820062d522\" (UID: \"062eea62-0f0f-45c4-a2e2-24820062d522\") " Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.180310 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/062eea62-0f0f-45c4-a2e2-24820062d522-logs" (OuterVolumeSpecName: "logs") pod "062eea62-0f0f-45c4-a2e2-24820062d522" (UID: "062eea62-0f0f-45c4-a2e2-24820062d522"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.188447 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/062eea62-0f0f-45c4-a2e2-24820062d522-kube-api-access-vg4x9" (OuterVolumeSpecName: "kube-api-access-vg4x9") pod "062eea62-0f0f-45c4-a2e2-24820062d522" (UID: "062eea62-0f0f-45c4-a2e2-24820062d522"). InnerVolumeSpecName "kube-api-access-vg4x9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.202658 4884 scope.go:117] "RemoveContainer" containerID="5eba1d389242792d3f6698b0d5b8aa88025a5f1c27ab83416aa209c596ce5ba5" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.209447 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "062eea62-0f0f-45c4-a2e2-24820062d522" (UID: "062eea62-0f0f-45c4-a2e2-24820062d522"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.242546 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "062eea62-0f0f-45c4-a2e2-24820062d522" (UID: "062eea62-0f0f-45c4-a2e2-24820062d522"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.270098 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-config-data" (OuterVolumeSpecName: "config-data") pod "062eea62-0f0f-45c4-a2e2-24820062d522" (UID: "062eea62-0f0f-45c4-a2e2-24820062d522"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.278922 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/062eea62-0f0f-45c4-a2e2-24820062d522-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.278950 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.278960 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vg4x9\" (UniqueName: \"kubernetes.io/projected/062eea62-0f0f-45c4-a2e2-24820062d522-kube-api-access-vg4x9\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.278970 4884 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.278992 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/062eea62-0f0f-45c4-a2e2-24820062d522-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.287225 4884 scope.go:117] "RemoveContainer" containerID="dcbd98665831ff5ed42a0debbf06d1ce2e3efeff9859ff024feeae70f3ac0952" Dec 02 01:59:33 crc kubenswrapper[4884]: E1202 01:59:33.287641 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcbd98665831ff5ed42a0debbf06d1ce2e3efeff9859ff024feeae70f3ac0952\": container with ID starting with dcbd98665831ff5ed42a0debbf06d1ce2e3efeff9859ff024feeae70f3ac0952 not found: ID does not exist" containerID="dcbd98665831ff5ed42a0debbf06d1ce2e3efeff9859ff024feeae70f3ac0952" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.287790 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcbd98665831ff5ed42a0debbf06d1ce2e3efeff9859ff024feeae70f3ac0952"} err="failed to get container status \"dcbd98665831ff5ed42a0debbf06d1ce2e3efeff9859ff024feeae70f3ac0952\": rpc error: code = NotFound desc = could not find container \"dcbd98665831ff5ed42a0debbf06d1ce2e3efeff9859ff024feeae70f3ac0952\": container with ID starting with dcbd98665831ff5ed42a0debbf06d1ce2e3efeff9859ff024feeae70f3ac0952 not found: ID does not exist" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.287816 4884 scope.go:117] "RemoveContainer" containerID="5eba1d389242792d3f6698b0d5b8aa88025a5f1c27ab83416aa209c596ce5ba5" Dec 02 01:59:33 crc kubenswrapper[4884]: E1202 01:59:33.288237 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5eba1d389242792d3f6698b0d5b8aa88025a5f1c27ab83416aa209c596ce5ba5\": container with ID starting with 5eba1d389242792d3f6698b0d5b8aa88025a5f1c27ab83416aa209c596ce5ba5 not found: ID does not exist" containerID="5eba1d389242792d3f6698b0d5b8aa88025a5f1c27ab83416aa209c596ce5ba5" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.288272 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5eba1d389242792d3f6698b0d5b8aa88025a5f1c27ab83416aa209c596ce5ba5"} err="failed to get container status \"5eba1d389242792d3f6698b0d5b8aa88025a5f1c27ab83416aa209c596ce5ba5\": rpc error: code = NotFound desc = could not find container \"5eba1d389242792d3f6698b0d5b8aa88025a5f1c27ab83416aa209c596ce5ba5\": container with ID starting with 5eba1d389242792d3f6698b0d5b8aa88025a5f1c27ab83416aa209c596ce5ba5 not found: ID does not exist" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.509508 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.518785 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.542833 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:59:33 crc kubenswrapper[4884]: E1202 01:59:33.543284 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="062eea62-0f0f-45c4-a2e2-24820062d522" containerName="nova-metadata-log" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.543301 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="062eea62-0f0f-45c4-a2e2-24820062d522" containerName="nova-metadata-log" Dec 02 01:59:33 crc kubenswrapper[4884]: E1202 01:59:33.543315 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1acb68f2-bed9-462f-82eb-244f84069402" containerName="nova-manage" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.543324 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="1acb68f2-bed9-462f-82eb-244f84069402" containerName="nova-manage" Dec 02 01:59:33 crc kubenswrapper[4884]: E1202 01:59:33.543339 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="062eea62-0f0f-45c4-a2e2-24820062d522" containerName="nova-metadata-metadata" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.543348 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="062eea62-0f0f-45c4-a2e2-24820062d522" containerName="nova-metadata-metadata" Dec 02 01:59:33 crc kubenswrapper[4884]: E1202 01:59:33.543391 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2686aabc-d1de-4d56-95db-0b919d055b3d" containerName="dnsmasq-dns" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.543399 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2686aabc-d1de-4d56-95db-0b919d055b3d" containerName="dnsmasq-dns" Dec 02 01:59:33 crc kubenswrapper[4884]: E1202 01:59:33.543415 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2686aabc-d1de-4d56-95db-0b919d055b3d" containerName="init" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.543421 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2686aabc-d1de-4d56-95db-0b919d055b3d" containerName="init" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.543844 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="1acb68f2-bed9-462f-82eb-244f84069402" containerName="nova-manage" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.543875 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="062eea62-0f0f-45c4-a2e2-24820062d522" containerName="nova-metadata-log" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.543887 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="2686aabc-d1de-4d56-95db-0b919d055b3d" containerName="dnsmasq-dns" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.543907 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="062eea62-0f0f-45c4-a2e2-24820062d522" containerName="nova-metadata-metadata" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.545016 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.547510 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.548051 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.568962 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.626474 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="062eea62-0f0f-45c4-a2e2-24820062d522" path="/var/lib/kubelet/pods/062eea62-0f0f-45c4-a2e2-24820062d522/volumes" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.686755 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e9beb88-6b94-45ad-ac2b-a86abd37c944-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9e9beb88-6b94-45ad-ac2b-a86abd37c944\") " pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.686834 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e9beb88-6b94-45ad-ac2b-a86abd37c944-config-data\") pod \"nova-metadata-0\" (UID: \"9e9beb88-6b94-45ad-ac2b-a86abd37c944\") " pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.686910 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e9beb88-6b94-45ad-ac2b-a86abd37c944-logs\") pod \"nova-metadata-0\" (UID: \"9e9beb88-6b94-45ad-ac2b-a86abd37c944\") " pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.686995 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r586q\" (UniqueName: \"kubernetes.io/projected/9e9beb88-6b94-45ad-ac2b-a86abd37c944-kube-api-access-r586q\") pod \"nova-metadata-0\" (UID: \"9e9beb88-6b94-45ad-ac2b-a86abd37c944\") " pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.687116 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e9beb88-6b94-45ad-ac2b-a86abd37c944-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9e9beb88-6b94-45ad-ac2b-a86abd37c944\") " pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.789141 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e9beb88-6b94-45ad-ac2b-a86abd37c944-logs\") pod \"nova-metadata-0\" (UID: \"9e9beb88-6b94-45ad-ac2b-a86abd37c944\") " pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.789213 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r586q\" (UniqueName: \"kubernetes.io/projected/9e9beb88-6b94-45ad-ac2b-a86abd37c944-kube-api-access-r586q\") pod \"nova-metadata-0\" (UID: \"9e9beb88-6b94-45ad-ac2b-a86abd37c944\") " pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.789264 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e9beb88-6b94-45ad-ac2b-a86abd37c944-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9e9beb88-6b94-45ad-ac2b-a86abd37c944\") " pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.789335 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e9beb88-6b94-45ad-ac2b-a86abd37c944-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9e9beb88-6b94-45ad-ac2b-a86abd37c944\") " pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.789359 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e9beb88-6b94-45ad-ac2b-a86abd37c944-config-data\") pod \"nova-metadata-0\" (UID: \"9e9beb88-6b94-45ad-ac2b-a86abd37c944\") " pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.789624 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e9beb88-6b94-45ad-ac2b-a86abd37c944-logs\") pod \"nova-metadata-0\" (UID: \"9e9beb88-6b94-45ad-ac2b-a86abd37c944\") " pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.792373 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e9beb88-6b94-45ad-ac2b-a86abd37c944-config-data\") pod \"nova-metadata-0\" (UID: \"9e9beb88-6b94-45ad-ac2b-a86abd37c944\") " pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.793562 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e9beb88-6b94-45ad-ac2b-a86abd37c944-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"9e9beb88-6b94-45ad-ac2b-a86abd37c944\") " pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.794142 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e9beb88-6b94-45ad-ac2b-a86abd37c944-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"9e9beb88-6b94-45ad-ac2b-a86abd37c944\") " pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.813089 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r586q\" (UniqueName: \"kubernetes.io/projected/9e9beb88-6b94-45ad-ac2b-a86abd37c944-kube-api-access-r586q\") pod \"nova-metadata-0\" (UID: \"9e9beb88-6b94-45ad-ac2b-a86abd37c944\") " pod="openstack/nova-metadata-0" Dec 02 01:59:33 crc kubenswrapper[4884]: I1202 01:59:33.873172 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.179169 4884 generic.go:334] "Generic (PLEG): container finished" podID="194a7ea1-a1f9-4868-9acc-628f909b9b11" containerID="67e389f9a1bbeff96e6d5b65b6cf85198346fa8edfe1cce938dc8a555607f6f1" exitCode=0 Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.179218 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"194a7ea1-a1f9-4868-9acc-628f909b9b11","Type":"ContainerDied","Data":"67e389f9a1bbeff96e6d5b65b6cf85198346fa8edfe1cce938dc8a555607f6f1"} Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.295412 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.401555 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-config-data\") pod \"194a7ea1-a1f9-4868-9acc-628f909b9b11\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.401642 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-combined-ca-bundle\") pod \"194a7ea1-a1f9-4868-9acc-628f909b9b11\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.401799 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/194a7ea1-a1f9-4868-9acc-628f909b9b11-logs\") pod \"194a7ea1-a1f9-4868-9acc-628f909b9b11\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.401879 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmdgp\" (UniqueName: \"kubernetes.io/projected/194a7ea1-a1f9-4868-9acc-628f909b9b11-kube-api-access-jmdgp\") pod \"194a7ea1-a1f9-4868-9acc-628f909b9b11\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.401932 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-public-tls-certs\") pod \"194a7ea1-a1f9-4868-9acc-628f909b9b11\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.401974 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-internal-tls-certs\") pod \"194a7ea1-a1f9-4868-9acc-628f909b9b11\" (UID: \"194a7ea1-a1f9-4868-9acc-628f909b9b11\") " Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.402464 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/194a7ea1-a1f9-4868-9acc-628f909b9b11-logs" (OuterVolumeSpecName: "logs") pod "194a7ea1-a1f9-4868-9acc-628f909b9b11" (UID: "194a7ea1-a1f9-4868-9acc-628f909b9b11"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.402836 4884 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/194a7ea1-a1f9-4868-9acc-628f909b9b11-logs\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.406193 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/194a7ea1-a1f9-4868-9acc-628f909b9b11-kube-api-access-jmdgp" (OuterVolumeSpecName: "kube-api-access-jmdgp") pod "194a7ea1-a1f9-4868-9acc-628f909b9b11" (UID: "194a7ea1-a1f9-4868-9acc-628f909b9b11"). InnerVolumeSpecName "kube-api-access-jmdgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.410821 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.427157 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "194a7ea1-a1f9-4868-9acc-628f909b9b11" (UID: "194a7ea1-a1f9-4868-9acc-628f909b9b11"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.436189 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-config-data" (OuterVolumeSpecName: "config-data") pod "194a7ea1-a1f9-4868-9acc-628f909b9b11" (UID: "194a7ea1-a1f9-4868-9acc-628f909b9b11"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.447668 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "194a7ea1-a1f9-4868-9acc-628f909b9b11" (UID: "194a7ea1-a1f9-4868-9acc-628f909b9b11"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.450098 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "194a7ea1-a1f9-4868-9acc-628f909b9b11" (UID: "194a7ea1-a1f9-4868-9acc-628f909b9b11"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:34 crc kubenswrapper[4884]: W1202 01:59:34.466485 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e9beb88_6b94_45ad_ac2b_a86abd37c944.slice/crio-1ebd680a476f311713bd238078158ccfc33ff5d47a8a3831ee798bc2ef8cc572 WatchSource:0}: Error finding container 1ebd680a476f311713bd238078158ccfc33ff5d47a8a3831ee798bc2ef8cc572: Status 404 returned error can't find the container with id 1ebd680a476f311713bd238078158ccfc33ff5d47a8a3831ee798bc2ef8cc572 Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.505155 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.505189 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.505203 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmdgp\" (UniqueName: \"kubernetes.io/projected/194a7ea1-a1f9-4868-9acc-628f909b9b11-kube-api-access-jmdgp\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.505211 4884 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.505220 4884 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/194a7ea1-a1f9-4868-9acc-628f909b9b11-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.879938 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.910277 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17a0f0d-1e28-4238-83f8-742b3b3df02c-config-data\") pod \"d17a0f0d-1e28-4238-83f8-742b3b3df02c\" (UID: \"d17a0f0d-1e28-4238-83f8-742b3b3df02c\") " Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.910391 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5rr8\" (UniqueName: \"kubernetes.io/projected/d17a0f0d-1e28-4238-83f8-742b3b3df02c-kube-api-access-d5rr8\") pod \"d17a0f0d-1e28-4238-83f8-742b3b3df02c\" (UID: \"d17a0f0d-1e28-4238-83f8-742b3b3df02c\") " Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.910696 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17a0f0d-1e28-4238-83f8-742b3b3df02c-combined-ca-bundle\") pod \"d17a0f0d-1e28-4238-83f8-742b3b3df02c\" (UID: \"d17a0f0d-1e28-4238-83f8-742b3b3df02c\") " Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.920232 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d17a0f0d-1e28-4238-83f8-742b3b3df02c-kube-api-access-d5rr8" (OuterVolumeSpecName: "kube-api-access-d5rr8") pod "d17a0f0d-1e28-4238-83f8-742b3b3df02c" (UID: "d17a0f0d-1e28-4238-83f8-742b3b3df02c"). InnerVolumeSpecName "kube-api-access-d5rr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.951941 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17a0f0d-1e28-4238-83f8-742b3b3df02c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d17a0f0d-1e28-4238-83f8-742b3b3df02c" (UID: "d17a0f0d-1e28-4238-83f8-742b3b3df02c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:34 crc kubenswrapper[4884]: I1202 01:59:34.957102 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17a0f0d-1e28-4238-83f8-742b3b3df02c-config-data" (OuterVolumeSpecName: "config-data") pod "d17a0f0d-1e28-4238-83f8-742b3b3df02c" (UID: "d17a0f0d-1e28-4238-83f8-742b3b3df02c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.012674 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17a0f0d-1e28-4238-83f8-742b3b3df02c-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.012709 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5rr8\" (UniqueName: \"kubernetes.io/projected/d17a0f0d-1e28-4238-83f8-742b3b3df02c-kube-api-access-d5rr8\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.012719 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17a0f0d-1e28-4238-83f8-742b3b3df02c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.193481 4884 generic.go:334] "Generic (PLEG): container finished" podID="d17a0f0d-1e28-4238-83f8-742b3b3df02c" containerID="6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd" exitCode=0 Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.193607 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.195070 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d17a0f0d-1e28-4238-83f8-742b3b3df02c","Type":"ContainerDied","Data":"6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd"} Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.195156 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d17a0f0d-1e28-4238-83f8-742b3b3df02c","Type":"ContainerDied","Data":"a5727aa6caf1ac333652dbbee5fc3dcf4b7a5dd14b0d84dfcc590576198bbf3f"} Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.195187 4884 scope.go:117] "RemoveContainer" containerID="6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.199314 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"194a7ea1-a1f9-4868-9acc-628f909b9b11","Type":"ContainerDied","Data":"e0e6ed378b30a307833cc137fb11a42135844e10557df57ee98e796c6ff5c6a2"} Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.199361 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.205206 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9e9beb88-6b94-45ad-ac2b-a86abd37c944","Type":"ContainerStarted","Data":"f388b0c993319476e3b11e7357d6a00ad8eb20e3b39c8f01eab86fe8253e85cc"} Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.205257 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9e9beb88-6b94-45ad-ac2b-a86abd37c944","Type":"ContainerStarted","Data":"b8e821d4ea3c183291594adcc48bca2ff8e3429e7f7c3f303c1c4ef7a8a25a9c"} Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.205279 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"9e9beb88-6b94-45ad-ac2b-a86abd37c944","Type":"ContainerStarted","Data":"1ebd680a476f311713bd238078158ccfc33ff5d47a8a3831ee798bc2ef8cc572"} Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.235529 4884 scope.go:117] "RemoveContainer" containerID="6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd" Dec 02 01:59:35 crc kubenswrapper[4884]: E1202 01:59:35.237510 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd\": container with ID starting with 6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd not found: ID does not exist" containerID="6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.237549 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd"} err="failed to get container status \"6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd\": rpc error: code = NotFound desc = could not find container \"6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd\": container with ID starting with 6e39f6976888c32cb85003be49c313170d347b95571863fb6a22cdd744a84acd not found: ID does not exist" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.237575 4884 scope.go:117] "RemoveContainer" containerID="67e389f9a1bbeff96e6d5b65b6cf85198346fa8edfe1cce938dc8a555607f6f1" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.237777 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.23773636 podStartE2EDuration="2.23773636s" podCreationTimestamp="2025-12-02 01:59:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:59:35.223285714 +0000 UTC m=+1391.899122618" watchObservedRunningTime="2025-12-02 01:59:35.23773636 +0000 UTC m=+1391.913573284" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.264821 4884 scope.go:117] "RemoveContainer" containerID="b8caabff5dc73ebcb399f862b4f0319e13bc01e5b8b668c981308df68c05fd5d" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.284815 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.292189 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.315782 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.324137 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.332205 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 01:59:35 crc kubenswrapper[4884]: E1202 01:59:35.332668 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="194a7ea1-a1f9-4868-9acc-628f909b9b11" containerName="nova-api-log" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.332687 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="194a7ea1-a1f9-4868-9acc-628f909b9b11" containerName="nova-api-log" Dec 02 01:59:35 crc kubenswrapper[4884]: E1202 01:59:35.332707 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="194a7ea1-a1f9-4868-9acc-628f909b9b11" containerName="nova-api-api" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.332713 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="194a7ea1-a1f9-4868-9acc-628f909b9b11" containerName="nova-api-api" Dec 02 01:59:35 crc kubenswrapper[4884]: E1202 01:59:35.332723 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17a0f0d-1e28-4238-83f8-742b3b3df02c" containerName="nova-scheduler-scheduler" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.332728 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17a0f0d-1e28-4238-83f8-742b3b3df02c" containerName="nova-scheduler-scheduler" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.332917 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d17a0f0d-1e28-4238-83f8-742b3b3df02c" containerName="nova-scheduler-scheduler" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.332938 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="194a7ea1-a1f9-4868-9acc-628f909b9b11" containerName="nova-api-log" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.332954 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="194a7ea1-a1f9-4868-9acc-628f909b9b11" containerName="nova-api-api" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.333637 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.336180 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.340463 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.342294 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.344047 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.344169 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.346115 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.355859 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.386552 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.422651 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.422697 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d8bv\" (UniqueName: \"kubernetes.io/projected/1a34c2cd-4194-4086-95a2-3d016faa05db-kube-api-access-5d8bv\") pod \"nova-scheduler-0\" (UID: \"1a34c2cd-4194-4086-95a2-3d016faa05db\") " pod="openstack/nova-scheduler-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.422716 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a34c2cd-4194-4086-95a2-3d016faa05db-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1a34c2cd-4194-4086-95a2-3d016faa05db\") " pod="openstack/nova-scheduler-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.422769 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-logs\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.422787 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-public-tls-certs\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.422863 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh79k\" (UniqueName: \"kubernetes.io/projected/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-kube-api-access-kh79k\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.422991 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.423049 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a34c2cd-4194-4086-95a2-3d016faa05db-config-data\") pod \"nova-scheduler-0\" (UID: \"1a34c2cd-4194-4086-95a2-3d016faa05db\") " pod="openstack/nova-scheduler-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.423102 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-config-data\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.525390 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a34c2cd-4194-4086-95a2-3d016faa05db-config-data\") pod \"nova-scheduler-0\" (UID: \"1a34c2cd-4194-4086-95a2-3d016faa05db\") " pod="openstack/nova-scheduler-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.525491 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-config-data\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.525598 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.525626 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d8bv\" (UniqueName: \"kubernetes.io/projected/1a34c2cd-4194-4086-95a2-3d016faa05db-kube-api-access-5d8bv\") pod \"nova-scheduler-0\" (UID: \"1a34c2cd-4194-4086-95a2-3d016faa05db\") " pod="openstack/nova-scheduler-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.525650 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a34c2cd-4194-4086-95a2-3d016faa05db-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1a34c2cd-4194-4086-95a2-3d016faa05db\") " pod="openstack/nova-scheduler-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.525694 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-logs\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.525718 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-public-tls-certs\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.525765 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh79k\" (UniqueName: \"kubernetes.io/projected/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-kube-api-access-kh79k\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.525795 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.530671 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-logs\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.530915 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-public-tls-certs\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.531500 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-internal-tls-certs\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.532124 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.532249 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a34c2cd-4194-4086-95a2-3d016faa05db-config-data\") pod \"nova-scheduler-0\" (UID: \"1a34c2cd-4194-4086-95a2-3d016faa05db\") " pod="openstack/nova-scheduler-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.532490 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a34c2cd-4194-4086-95a2-3d016faa05db-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1a34c2cd-4194-4086-95a2-3d016faa05db\") " pod="openstack/nova-scheduler-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.534431 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-config-data\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.548487 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d8bv\" (UniqueName: \"kubernetes.io/projected/1a34c2cd-4194-4086-95a2-3d016faa05db-kube-api-access-5d8bv\") pod \"nova-scheduler-0\" (UID: \"1a34c2cd-4194-4086-95a2-3d016faa05db\") " pod="openstack/nova-scheduler-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.549692 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh79k\" (UniqueName: \"kubernetes.io/projected/0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d-kube-api-access-kh79k\") pod \"nova-api-0\" (UID: \"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d\") " pod="openstack/nova-api-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.628122 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="194a7ea1-a1f9-4868-9acc-628f909b9b11" path="/var/lib/kubelet/pods/194a7ea1-a1f9-4868-9acc-628f909b9b11/volumes" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.628714 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d17a0f0d-1e28-4238-83f8-742b3b3df02c" path="/var/lib/kubelet/pods/d17a0f0d-1e28-4238-83f8-742b3b3df02c/volumes" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.653869 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 02 01:59:35 crc kubenswrapper[4884]: I1202 01:59:35.664751 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 02 01:59:36 crc kubenswrapper[4884]: I1202 01:59:36.143304 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 02 01:59:36 crc kubenswrapper[4884]: I1202 01:59:36.229787 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1a34c2cd-4194-4086-95a2-3d016faa05db","Type":"ContainerStarted","Data":"44ab9c036459c18ab8d68387aca1d0f3722341d4a6d7ca44e7c798698d865f4f"} Dec 02 01:59:36 crc kubenswrapper[4884]: I1202 01:59:36.260810 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 02 01:59:37 crc kubenswrapper[4884]: I1202 01:59:37.249819 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d","Type":"ContainerStarted","Data":"cc5c90088aeb2228099821d568a45b422e6edc84d30eb37c8a59b29109ca922d"} Dec 02 01:59:37 crc kubenswrapper[4884]: I1202 01:59:37.250148 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d","Type":"ContainerStarted","Data":"6a3192ff4773c1503375111dadb32b6e3c4a9f5be100923e80cafc454fba198b"} Dec 02 01:59:37 crc kubenswrapper[4884]: I1202 01:59:37.250165 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d","Type":"ContainerStarted","Data":"3cd6b120add9955e28c81492216fbeecc3924e8dd42f0e8b00c7ce223e1ccc9b"} Dec 02 01:59:37 crc kubenswrapper[4884]: I1202 01:59:37.254019 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1a34c2cd-4194-4086-95a2-3d016faa05db","Type":"ContainerStarted","Data":"c44f5455122a043153f005f25a28f43303cd13b4fbcffe25f0ee936bc52a5741"} Dec 02 01:59:37 crc kubenswrapper[4884]: I1202 01:59:37.283939 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.283915762 podStartE2EDuration="2.283915762s" podCreationTimestamp="2025-12-02 01:59:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:59:37.277826541 +0000 UTC m=+1393.953663496" watchObservedRunningTime="2025-12-02 01:59:37.283915762 +0000 UTC m=+1393.959752686" Dec 02 01:59:37 crc kubenswrapper[4884]: I1202 01:59:37.893951 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="062eea62-0f0f-45c4-a2e2-24820062d522" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 02 01:59:37 crc kubenswrapper[4884]: I1202 01:59:37.894007 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="062eea62-0f0f-45c4-a2e2-24820062d522" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 02 01:59:38 crc kubenswrapper[4884]: I1202 01:59:38.874185 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 01:59:38 crc kubenswrapper[4884]: I1202 01:59:38.874592 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 02 01:59:40 crc kubenswrapper[4884]: I1202 01:59:40.654583 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 02 01:59:43 crc kubenswrapper[4884]: I1202 01:59:43.874354 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 02 01:59:43 crc kubenswrapper[4884]: I1202 01:59:43.874792 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 02 01:59:44 crc kubenswrapper[4884]: I1202 01:59:44.886978 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9e9beb88-6b94-45ad-ac2b-a86abd37c944" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.219:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 02 01:59:44 crc kubenswrapper[4884]: I1202 01:59:44.887411 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="9e9beb88-6b94-45ad-ac2b-a86abd37c944" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.219:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 01:59:45 crc kubenswrapper[4884]: I1202 01:59:45.654101 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 02 01:59:45 crc kubenswrapper[4884]: I1202 01:59:45.665679 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 01:59:45 crc kubenswrapper[4884]: I1202 01:59:45.665716 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 02 01:59:45 crc kubenswrapper[4884]: I1202 01:59:45.687251 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 02 01:59:45 crc kubenswrapper[4884]: I1202 01:59:45.713263 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=10.713245603 podStartE2EDuration="10.713245603s" podCreationTimestamp="2025-12-02 01:59:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 01:59:37.311192883 +0000 UTC m=+1393.987029777" watchObservedRunningTime="2025-12-02 01:59:45.713245603 +0000 UTC m=+1402.389082487" Dec 02 01:59:46 crc kubenswrapper[4884]: I1202 01:59:46.427631 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 02 01:59:46 crc kubenswrapper[4884]: I1202 01:59:46.683876 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.221:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 01:59:46 crc kubenswrapper[4884]: I1202 01:59:46.683932 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.221:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 02 01:59:46 crc kubenswrapper[4884]: I1202 01:59:46.971380 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 01:59:46 crc kubenswrapper[4884]: I1202 01:59:46.971450 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 01:59:47 crc kubenswrapper[4884]: I1202 01:59:47.363190 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 02 01:59:53 crc kubenswrapper[4884]: I1202 01:59:53.885683 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 02 01:59:53 crc kubenswrapper[4884]: I1202 01:59:53.899128 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 02 01:59:53 crc kubenswrapper[4884]: I1202 01:59:53.901604 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 02 01:59:54 crc kubenswrapper[4884]: I1202 01:59:54.481303 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 02 01:59:55 crc kubenswrapper[4884]: I1202 01:59:55.674426 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 02 01:59:55 crc kubenswrapper[4884]: I1202 01:59:55.676377 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 02 01:59:55 crc kubenswrapper[4884]: I1202 01:59:55.681925 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 02 01:59:55 crc kubenswrapper[4884]: I1202 01:59:55.689257 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 02 01:59:56 crc kubenswrapper[4884]: I1202 01:59:56.506011 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 02 01:59:56 crc kubenswrapper[4884]: I1202 01:59:56.518972 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 02 02:00:00 crc kubenswrapper[4884]: I1202 02:00:00.174493 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn"] Dec 02 02:00:00 crc kubenswrapper[4884]: I1202 02:00:00.176730 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" Dec 02 02:00:00 crc kubenswrapper[4884]: I1202 02:00:00.179899 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 02:00:00 crc kubenswrapper[4884]: I1202 02:00:00.180197 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 02:00:00 crc kubenswrapper[4884]: I1202 02:00:00.191520 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn"] Dec 02 02:00:00 crc kubenswrapper[4884]: I1202 02:00:00.308335 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5869070c-4b43-4ecc-b6a0-971cd0686284-config-volume\") pod \"collect-profiles-29410680-qmndn\" (UID: \"5869070c-4b43-4ecc-b6a0-971cd0686284\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" Dec 02 02:00:00 crc kubenswrapper[4884]: I1202 02:00:00.308419 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5869070c-4b43-4ecc-b6a0-971cd0686284-secret-volume\") pod \"collect-profiles-29410680-qmndn\" (UID: \"5869070c-4b43-4ecc-b6a0-971cd0686284\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" Dec 02 02:00:00 crc kubenswrapper[4884]: I1202 02:00:00.308466 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnds7\" (UniqueName: \"kubernetes.io/projected/5869070c-4b43-4ecc-b6a0-971cd0686284-kube-api-access-gnds7\") pod \"collect-profiles-29410680-qmndn\" (UID: \"5869070c-4b43-4ecc-b6a0-971cd0686284\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" Dec 02 02:00:00 crc kubenswrapper[4884]: I1202 02:00:00.411016 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5869070c-4b43-4ecc-b6a0-971cd0686284-config-volume\") pod \"collect-profiles-29410680-qmndn\" (UID: \"5869070c-4b43-4ecc-b6a0-971cd0686284\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" Dec 02 02:00:00 crc kubenswrapper[4884]: I1202 02:00:00.411084 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5869070c-4b43-4ecc-b6a0-971cd0686284-secret-volume\") pod \"collect-profiles-29410680-qmndn\" (UID: \"5869070c-4b43-4ecc-b6a0-971cd0686284\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" Dec 02 02:00:00 crc kubenswrapper[4884]: I1202 02:00:00.411119 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnds7\" (UniqueName: \"kubernetes.io/projected/5869070c-4b43-4ecc-b6a0-971cd0686284-kube-api-access-gnds7\") pod \"collect-profiles-29410680-qmndn\" (UID: \"5869070c-4b43-4ecc-b6a0-971cd0686284\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" Dec 02 02:00:00 crc kubenswrapper[4884]: I1202 02:00:00.412107 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5869070c-4b43-4ecc-b6a0-971cd0686284-config-volume\") pod \"collect-profiles-29410680-qmndn\" (UID: \"5869070c-4b43-4ecc-b6a0-971cd0686284\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" Dec 02 02:00:00 crc kubenswrapper[4884]: I1202 02:00:00.419148 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5869070c-4b43-4ecc-b6a0-971cd0686284-secret-volume\") pod \"collect-profiles-29410680-qmndn\" (UID: \"5869070c-4b43-4ecc-b6a0-971cd0686284\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" Dec 02 02:00:00 crc kubenswrapper[4884]: I1202 02:00:00.431576 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnds7\" (UniqueName: \"kubernetes.io/projected/5869070c-4b43-4ecc-b6a0-971cd0686284-kube-api-access-gnds7\") pod \"collect-profiles-29410680-qmndn\" (UID: \"5869070c-4b43-4ecc-b6a0-971cd0686284\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" Dec 02 02:00:00 crc kubenswrapper[4884]: I1202 02:00:00.522233 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" Dec 02 02:00:01 crc kubenswrapper[4884]: W1202 02:00:00.973572 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5869070c_4b43_4ecc_b6a0_971cd0686284.slice/crio-6623b0d37857efede4505ed65320d645e57a2086d68bf0e2b941dc13bd033946 WatchSource:0}: Error finding container 6623b0d37857efede4505ed65320d645e57a2086d68bf0e2b941dc13bd033946: Status 404 returned error can't find the container with id 6623b0d37857efede4505ed65320d645e57a2086d68bf0e2b941dc13bd033946 Dec 02 02:00:01 crc kubenswrapper[4884]: I1202 02:00:00.980186 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn"] Dec 02 02:00:01 crc kubenswrapper[4884]: I1202 02:00:01.572566 4884 generic.go:334] "Generic (PLEG): container finished" podID="5869070c-4b43-4ecc-b6a0-971cd0686284" containerID="ce10549b269f041591f6b05a9a730ea509a7dbdeeb9780199b39e6516401a936" exitCode=0 Dec 02 02:00:01 crc kubenswrapper[4884]: I1202 02:00:01.572802 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" event={"ID":"5869070c-4b43-4ecc-b6a0-971cd0686284","Type":"ContainerDied","Data":"ce10549b269f041591f6b05a9a730ea509a7dbdeeb9780199b39e6516401a936"} Dec 02 02:00:01 crc kubenswrapper[4884]: I1202 02:00:01.573077 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" event={"ID":"5869070c-4b43-4ecc-b6a0-971cd0686284","Type":"ContainerStarted","Data":"6623b0d37857efede4505ed65320d645e57a2086d68bf0e2b941dc13bd033946"} Dec 02 02:00:02 crc kubenswrapper[4884]: I1202 02:00:02.993421 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" Dec 02 02:00:03 crc kubenswrapper[4884]: I1202 02:00:03.080416 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5869070c-4b43-4ecc-b6a0-971cd0686284-config-volume\") pod \"5869070c-4b43-4ecc-b6a0-971cd0686284\" (UID: \"5869070c-4b43-4ecc-b6a0-971cd0686284\") " Dec 02 02:00:03 crc kubenswrapper[4884]: I1202 02:00:03.080566 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5869070c-4b43-4ecc-b6a0-971cd0686284-secret-volume\") pod \"5869070c-4b43-4ecc-b6a0-971cd0686284\" (UID: \"5869070c-4b43-4ecc-b6a0-971cd0686284\") " Dec 02 02:00:03 crc kubenswrapper[4884]: I1202 02:00:03.080699 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnds7\" (UniqueName: \"kubernetes.io/projected/5869070c-4b43-4ecc-b6a0-971cd0686284-kube-api-access-gnds7\") pod \"5869070c-4b43-4ecc-b6a0-971cd0686284\" (UID: \"5869070c-4b43-4ecc-b6a0-971cd0686284\") " Dec 02 02:00:03 crc kubenswrapper[4884]: I1202 02:00:03.081554 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5869070c-4b43-4ecc-b6a0-971cd0686284-config-volume" (OuterVolumeSpecName: "config-volume") pod "5869070c-4b43-4ecc-b6a0-971cd0686284" (UID: "5869070c-4b43-4ecc-b6a0-971cd0686284"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:03 crc kubenswrapper[4884]: I1202 02:00:03.090429 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5869070c-4b43-4ecc-b6a0-971cd0686284-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5869070c-4b43-4ecc-b6a0-971cd0686284" (UID: "5869070c-4b43-4ecc-b6a0-971cd0686284"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:00:03 crc kubenswrapper[4884]: I1202 02:00:03.090443 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5869070c-4b43-4ecc-b6a0-971cd0686284-kube-api-access-gnds7" (OuterVolumeSpecName: "kube-api-access-gnds7") pod "5869070c-4b43-4ecc-b6a0-971cd0686284" (UID: "5869070c-4b43-4ecc-b6a0-971cd0686284"). InnerVolumeSpecName "kube-api-access-gnds7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:00:03 crc kubenswrapper[4884]: I1202 02:00:03.183666 4884 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5869070c-4b43-4ecc-b6a0-971cd0686284-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:03 crc kubenswrapper[4884]: I1202 02:00:03.183727 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnds7\" (UniqueName: \"kubernetes.io/projected/5869070c-4b43-4ecc-b6a0-971cd0686284-kube-api-access-gnds7\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:03 crc kubenswrapper[4884]: I1202 02:00:03.183774 4884 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5869070c-4b43-4ecc-b6a0-971cd0686284-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:03 crc kubenswrapper[4884]: I1202 02:00:03.601239 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" event={"ID":"5869070c-4b43-4ecc-b6a0-971cd0686284","Type":"ContainerDied","Data":"6623b0d37857efede4505ed65320d645e57a2086d68bf0e2b941dc13bd033946"} Dec 02 02:00:03 crc kubenswrapper[4884]: I1202 02:00:03.601299 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6623b0d37857efede4505ed65320d645e57a2086d68bf0e2b941dc13bd033946" Dec 02 02:00:03 crc kubenswrapper[4884]: I1202 02:00:03.601374 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn" Dec 02 02:00:04 crc kubenswrapper[4884]: I1202 02:00:04.607440 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 02:00:05 crc kubenswrapper[4884]: I1202 02:00:05.415273 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 02:00:08 crc kubenswrapper[4884]: I1202 02:00:08.673365 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="f32d61be-992e-4a09-9663-33d3d59779e9" containerName="rabbitmq" containerID="cri-o://667241041ac23c701da96842218f42323b58836f1801238ed1c63c9a447b69d9" gracePeriod=604796 Dec 02 02:00:09 crc kubenswrapper[4884]: I1202 02:00:09.311870 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="27182112-178e-415c-9af9-f7fbf58d8120" containerName="rabbitmq" containerID="cri-o://37cfb917f66312e86b6bacb7e41b56101104db52b4ae5ab3ec242ca5f852baf0" gracePeriod=604797 Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.281122 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.363282 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"f32d61be-992e-4a09-9663-33d3d59779e9\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.363407 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-plugins\") pod \"f32d61be-992e-4a09-9663-33d3d59779e9\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.363469 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-server-conf\") pod \"f32d61be-992e-4a09-9663-33d3d59779e9\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.363548 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-plugins-conf\") pod \"f32d61be-992e-4a09-9663-33d3d59779e9\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.363607 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-confd\") pod \"f32d61be-992e-4a09-9663-33d3d59779e9\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.363652 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f32d61be-992e-4a09-9663-33d3d59779e9-pod-info\") pod \"f32d61be-992e-4a09-9663-33d3d59779e9\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.363700 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-tls\") pod \"f32d61be-992e-4a09-9663-33d3d59779e9\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.363829 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm5x9\" (UniqueName: \"kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-kube-api-access-sm5x9\") pod \"f32d61be-992e-4a09-9663-33d3d59779e9\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.363882 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-erlang-cookie\") pod \"f32d61be-992e-4a09-9663-33d3d59779e9\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.363953 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-config-data\") pod \"f32d61be-992e-4a09-9663-33d3d59779e9\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.363999 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f32d61be-992e-4a09-9663-33d3d59779e9-erlang-cookie-secret\") pod \"f32d61be-992e-4a09-9663-33d3d59779e9\" (UID: \"f32d61be-992e-4a09-9663-33d3d59779e9\") " Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.366527 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "f32d61be-992e-4a09-9663-33d3d59779e9" (UID: "f32d61be-992e-4a09-9663-33d3d59779e9"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.368989 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "f32d61be-992e-4a09-9663-33d3d59779e9" (UID: "f32d61be-992e-4a09-9663-33d3d59779e9"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.369869 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "f32d61be-992e-4a09-9663-33d3d59779e9" (UID: "f32d61be-992e-4a09-9663-33d3d59779e9"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.375337 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "f32d61be-992e-4a09-9663-33d3d59779e9" (UID: "f32d61be-992e-4a09-9663-33d3d59779e9"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.375806 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f32d61be-992e-4a09-9663-33d3d59779e9-pod-info" (OuterVolumeSpecName: "pod-info") pod "f32d61be-992e-4a09-9663-33d3d59779e9" (UID: "f32d61be-992e-4a09-9663-33d3d59779e9"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.379836 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-kube-api-access-sm5x9" (OuterVolumeSpecName: "kube-api-access-sm5x9") pod "f32d61be-992e-4a09-9663-33d3d59779e9" (UID: "f32d61be-992e-4a09-9663-33d3d59779e9"). InnerVolumeSpecName "kube-api-access-sm5x9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.382761 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f32d61be-992e-4a09-9663-33d3d59779e9-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "f32d61be-992e-4a09-9663-33d3d59779e9" (UID: "f32d61be-992e-4a09-9663-33d3d59779e9"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.388507 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "f32d61be-992e-4a09-9663-33d3d59779e9" (UID: "f32d61be-992e-4a09-9663-33d3d59779e9"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.442786 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-config-data" (OuterVolumeSpecName: "config-data") pod "f32d61be-992e-4a09-9663-33d3d59779e9" (UID: "f32d61be-992e-4a09-9663-33d3d59779e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.464184 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-server-conf" (OuterVolumeSpecName: "server-conf") pod "f32d61be-992e-4a09-9663-33d3d59779e9" (UID: "f32d61be-992e-4a09-9663-33d3d59779e9"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.468232 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm5x9\" (UniqueName: \"kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-kube-api-access-sm5x9\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.468262 4884 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.468272 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.468300 4884 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f32d61be-992e-4a09-9663-33d3d59779e9-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.468323 4884 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.468333 4884 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.468341 4884 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-server-conf\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.468349 4884 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f32d61be-992e-4a09-9663-33d3d59779e9-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.468664 4884 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f32d61be-992e-4a09-9663-33d3d59779e9-pod-info\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.468676 4884 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.495193 4884 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.542888 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "f32d61be-992e-4a09-9663-33d3d59779e9" (UID: "f32d61be-992e-4a09-9663-33d3d59779e9"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.570631 4884 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.570669 4884 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f32d61be-992e-4a09-9663-33d3d59779e9-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.764678 4884 generic.go:334] "Generic (PLEG): container finished" podID="f32d61be-992e-4a09-9663-33d3d59779e9" containerID="667241041ac23c701da96842218f42323b58836f1801238ed1c63c9a447b69d9" exitCode=0 Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.765170 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f32d61be-992e-4a09-9663-33d3d59779e9","Type":"ContainerDied","Data":"667241041ac23c701da96842218f42323b58836f1801238ed1c63c9a447b69d9"} Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.765198 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"f32d61be-992e-4a09-9663-33d3d59779e9","Type":"ContainerDied","Data":"cd0bcd038f0ce6837b08f6f928b5bd49d8ea3667fb693985e21a97472f5f63b2"} Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.765215 4884 scope.go:117] "RemoveContainer" containerID="667241041ac23c701da96842218f42323b58836f1801238ed1c63c9a447b69d9" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.765221 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.781290 4884 generic.go:334] "Generic (PLEG): container finished" podID="27182112-178e-415c-9af9-f7fbf58d8120" containerID="37cfb917f66312e86b6bacb7e41b56101104db52b4ae5ab3ec242ca5f852baf0" exitCode=0 Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.781370 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"27182112-178e-415c-9af9-f7fbf58d8120","Type":"ContainerDied","Data":"37cfb917f66312e86b6bacb7e41b56101104db52b4ae5ab3ec242ca5f852baf0"} Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.850881 4884 scope.go:117] "RemoveContainer" containerID="39010f45dce6a3ffe62ec61181fd5ae24a376cc18ac028cb8715cf1211681fc2" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.859015 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.879034 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.896290 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 02:00:15 crc kubenswrapper[4884]: E1202 02:00:15.896809 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f32d61be-992e-4a09-9663-33d3d59779e9" containerName="setup-container" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.896827 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f32d61be-992e-4a09-9663-33d3d59779e9" containerName="setup-container" Dec 02 02:00:15 crc kubenswrapper[4884]: E1202 02:00:15.896843 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f32d61be-992e-4a09-9663-33d3d59779e9" containerName="rabbitmq" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.896849 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f32d61be-992e-4a09-9663-33d3d59779e9" containerName="rabbitmq" Dec 02 02:00:15 crc kubenswrapper[4884]: E1202 02:00:15.896873 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5869070c-4b43-4ecc-b6a0-971cd0686284" containerName="collect-profiles" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.896879 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="5869070c-4b43-4ecc-b6a0-971cd0686284" containerName="collect-profiles" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.897045 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="f32d61be-992e-4a09-9663-33d3d59779e9" containerName="rabbitmq" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.897075 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="5869070c-4b43-4ecc-b6a0-971cd0686284" containerName="collect-profiles" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.898116 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.901344 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.901503 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.901663 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.902516 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-98v8z" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.902700 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.902906 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.903219 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.907675 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.917042 4884 scope.go:117] "RemoveContainer" containerID="667241041ac23c701da96842218f42323b58836f1801238ed1c63c9a447b69d9" Dec 02 02:00:15 crc kubenswrapper[4884]: E1202 02:00:15.929839 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"667241041ac23c701da96842218f42323b58836f1801238ed1c63c9a447b69d9\": container with ID starting with 667241041ac23c701da96842218f42323b58836f1801238ed1c63c9a447b69d9 not found: ID does not exist" containerID="667241041ac23c701da96842218f42323b58836f1801238ed1c63c9a447b69d9" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.930081 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"667241041ac23c701da96842218f42323b58836f1801238ed1c63c9a447b69d9"} err="failed to get container status \"667241041ac23c701da96842218f42323b58836f1801238ed1c63c9a447b69d9\": rpc error: code = NotFound desc = could not find container \"667241041ac23c701da96842218f42323b58836f1801238ed1c63c9a447b69d9\": container with ID starting with 667241041ac23c701da96842218f42323b58836f1801238ed1c63c9a447b69d9 not found: ID does not exist" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.930117 4884 scope.go:117] "RemoveContainer" containerID="39010f45dce6a3ffe62ec61181fd5ae24a376cc18ac028cb8715cf1211681fc2" Dec 02 02:00:15 crc kubenswrapper[4884]: E1202 02:00:15.930462 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39010f45dce6a3ffe62ec61181fd5ae24a376cc18ac028cb8715cf1211681fc2\": container with ID starting with 39010f45dce6a3ffe62ec61181fd5ae24a376cc18ac028cb8715cf1211681fc2 not found: ID does not exist" containerID="39010f45dce6a3ffe62ec61181fd5ae24a376cc18ac028cb8715cf1211681fc2" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.930501 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39010f45dce6a3ffe62ec61181fd5ae24a376cc18ac028cb8715cf1211681fc2"} err="failed to get container status \"39010f45dce6a3ffe62ec61181fd5ae24a376cc18ac028cb8715cf1211681fc2\": rpc error: code = NotFound desc = could not find container \"39010f45dce6a3ffe62ec61181fd5ae24a376cc18ac028cb8715cf1211681fc2\": container with ID starting with 39010f45dce6a3ffe62ec61181fd5ae24a376cc18ac028cb8715cf1211681fc2 not found: ID does not exist" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.937817 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.985352 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.985407 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.985431 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t5wb\" (UniqueName: \"kubernetes.io/projected/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-kube-api-access-8t5wb\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.985449 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.985554 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.985573 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.985593 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.985636 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.985658 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.985674 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:15 crc kubenswrapper[4884]: I1202 02:00:15.985717 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-config-data\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.087202 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-confd\") pod \"27182112-178e-415c-9af9-f7fbf58d8120\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.087295 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kk59\" (UniqueName: \"kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-kube-api-access-7kk59\") pod \"27182112-178e-415c-9af9-f7fbf58d8120\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.087329 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-config-data\") pod \"27182112-178e-415c-9af9-f7fbf58d8120\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.087440 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-server-conf\") pod \"27182112-178e-415c-9af9-f7fbf58d8120\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.087457 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/27182112-178e-415c-9af9-f7fbf58d8120-pod-info\") pod \"27182112-178e-415c-9af9-f7fbf58d8120\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.087477 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-tls\") pod \"27182112-178e-415c-9af9-f7fbf58d8120\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.087510 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-erlang-cookie\") pod \"27182112-178e-415c-9af9-f7fbf58d8120\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.087527 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-plugins\") pod \"27182112-178e-415c-9af9-f7fbf58d8120\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.087563 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"27182112-178e-415c-9af9-f7fbf58d8120\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.087579 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-plugins-conf\") pod \"27182112-178e-415c-9af9-f7fbf58d8120\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.087662 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/27182112-178e-415c-9af9-f7fbf58d8120-erlang-cookie-secret\") pod \"27182112-178e-415c-9af9-f7fbf58d8120\" (UID: \"27182112-178e-415c-9af9-f7fbf58d8120\") " Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.087958 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.088004 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.088030 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t5wb\" (UniqueName: \"kubernetes.io/projected/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-kube-api-access-8t5wb\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.088046 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.088097 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.088117 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.088137 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.088183 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.088205 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.088221 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.088263 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-config-data\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.088535 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "27182112-178e-415c-9af9-f7fbf58d8120" (UID: "27182112-178e-415c-9af9-f7fbf58d8120"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.088529 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "27182112-178e-415c-9af9-f7fbf58d8120" (UID: "27182112-178e-415c-9af9-f7fbf58d8120"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.089146 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-config-data\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.089322 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "27182112-178e-415c-9af9-f7fbf58d8120" (UID: "27182112-178e-415c-9af9-f7fbf58d8120"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.092565 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27182112-178e-415c-9af9-f7fbf58d8120-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "27182112-178e-415c-9af9-f7fbf58d8120" (UID: "27182112-178e-415c-9af9-f7fbf58d8120"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.093018 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-kube-api-access-7kk59" (OuterVolumeSpecName: "kube-api-access-7kk59") pod "27182112-178e-415c-9af9-f7fbf58d8120" (UID: "27182112-178e-415c-9af9-f7fbf58d8120"). InnerVolumeSpecName "kube-api-access-7kk59". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.093099 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "27182112-178e-415c-9af9-f7fbf58d8120" (UID: "27182112-178e-415c-9af9-f7fbf58d8120"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.094461 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.095307 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.095610 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.096237 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.096282 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "27182112-178e-415c-9af9-f7fbf58d8120" (UID: "27182112-178e-415c-9af9-f7fbf58d8120"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.096449 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.099253 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.112502 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.114252 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/27182112-178e-415c-9af9-f7fbf58d8120-pod-info" (OuterVolumeSpecName: "pod-info") pod "27182112-178e-415c-9af9-f7fbf58d8120" (UID: "27182112-178e-415c-9af9-f7fbf58d8120"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.118661 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t5wb\" (UniqueName: \"kubernetes.io/projected/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-kube-api-access-8t5wb\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.118658 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.123554 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.139015 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-config-data" (OuterVolumeSpecName: "config-data") pod "27182112-178e-415c-9af9-f7fbf58d8120" (UID: "27182112-178e-415c-9af9-f7fbf58d8120"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.181573 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-server-0\" (UID: \"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c\") " pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.187613 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-server-conf" (OuterVolumeSpecName: "server-conf") pod "27182112-178e-415c-9af9-f7fbf58d8120" (UID: "27182112-178e-415c-9af9-f7fbf58d8120"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.191264 4884 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-server-conf\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.191290 4884 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/27182112-178e-415c-9af9-f7fbf58d8120-pod-info\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.191299 4884 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.191310 4884 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.191319 4884 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.191343 4884 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.191354 4884 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.191362 4884 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/27182112-178e-415c-9af9-f7fbf58d8120-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.191371 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kk59\" (UniqueName: \"kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-kube-api-access-7kk59\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.191381 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/27182112-178e-415c-9af9-f7fbf58d8120-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.225862 4884 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.231258 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.257355 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "27182112-178e-415c-9af9-f7fbf58d8120" (UID: "27182112-178e-415c-9af9-f7fbf58d8120"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.292868 4884 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/27182112-178e-415c-9af9-f7fbf58d8120-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.293335 4884 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.704031 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 02 02:00:16 crc kubenswrapper[4884]: W1202 02:00:16.721091 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ed44e1a_89f9_4e26_9a9c_f83ee9b0f50c.slice/crio-8a84be134040c365f2d8e2967875439ba0d3d83f0d19cadbf2ca9f3552f32495 WatchSource:0}: Error finding container 8a84be134040c365f2d8e2967875439ba0d3d83f0d19cadbf2ca9f3552f32495: Status 404 returned error can't find the container with id 8a84be134040c365f2d8e2967875439ba0d3d83f0d19cadbf2ca9f3552f32495 Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.799292 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"27182112-178e-415c-9af9-f7fbf58d8120","Type":"ContainerDied","Data":"448febc64be0ef9b1b26959f87b8bfd21fe6ef4b9a57b1cc6c5804bb801e22b2"} Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.799785 4884 scope.go:117] "RemoveContainer" containerID="37cfb917f66312e86b6bacb7e41b56101104db52b4ae5ab3ec242ca5f852baf0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.799347 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.805395 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c","Type":"ContainerStarted","Data":"8a84be134040c365f2d8e2967875439ba0d3d83f0d19cadbf2ca9f3552f32495"} Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.857325 4884 scope.go:117] "RemoveContainer" containerID="b9dfb8cc8de08b7d246a2396788751343d208c51ed83bbee65b5010ac935f251" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.864899 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.880227 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.891599 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 02:00:16 crc kubenswrapper[4884]: E1202 02:00:16.892018 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27182112-178e-415c-9af9-f7fbf58d8120" containerName="rabbitmq" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.892035 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="27182112-178e-415c-9af9-f7fbf58d8120" containerName="rabbitmq" Dec 02 02:00:16 crc kubenswrapper[4884]: E1202 02:00:16.892073 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27182112-178e-415c-9af9-f7fbf58d8120" containerName="setup-container" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.892080 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="27182112-178e-415c-9af9-f7fbf58d8120" containerName="setup-container" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.892247 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="27182112-178e-415c-9af9-f7fbf58d8120" containerName="rabbitmq" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.893541 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.896802 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.897055 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-qhmqn" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.897328 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.897455 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.897762 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.897878 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.897980 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.915824 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.971866 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:00:16 crc kubenswrapper[4884]: I1202 02:00:16.971962 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.018561 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22n2c\" (UniqueName: \"kubernetes.io/projected/962f0d62-e23b-437e-ac98-aea141142a36-kube-api-access-22n2c\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.018632 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/962f0d62-e23b-437e-ac98-aea141142a36-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.018660 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/962f0d62-e23b-437e-ac98-aea141142a36-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.018694 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/962f0d62-e23b-437e-ac98-aea141142a36-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.018717 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/962f0d62-e23b-437e-ac98-aea141142a36-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.018917 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/962f0d62-e23b-437e-ac98-aea141142a36-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.019204 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/962f0d62-e23b-437e-ac98-aea141142a36-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.019328 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/962f0d62-e23b-437e-ac98-aea141142a36-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.019385 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.019567 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/962f0d62-e23b-437e-ac98-aea141142a36-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.019608 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/962f0d62-e23b-437e-ac98-aea141142a36-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.121385 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/962f0d62-e23b-437e-ac98-aea141142a36-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.121829 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.121930 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/962f0d62-e23b-437e-ac98-aea141142a36-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.122102 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/962f0d62-e23b-437e-ac98-aea141142a36-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.122188 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/962f0d62-e23b-437e-ac98-aea141142a36-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.122136 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.122318 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22n2c\" (UniqueName: \"kubernetes.io/projected/962f0d62-e23b-437e-ac98-aea141142a36-kube-api-access-22n2c\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.122396 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/962f0d62-e23b-437e-ac98-aea141142a36-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.122476 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/962f0d62-e23b-437e-ac98-aea141142a36-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.122562 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/962f0d62-e23b-437e-ac98-aea141142a36-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.122636 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/962f0d62-e23b-437e-ac98-aea141142a36-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.122723 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/962f0d62-e23b-437e-ac98-aea141142a36-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.122905 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/962f0d62-e23b-437e-ac98-aea141142a36-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.123343 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/962f0d62-e23b-437e-ac98-aea141142a36-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.123637 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/962f0d62-e23b-437e-ac98-aea141142a36-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.123995 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/962f0d62-e23b-437e-ac98-aea141142a36-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.124566 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/962f0d62-e23b-437e-ac98-aea141142a36-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.129425 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/962f0d62-e23b-437e-ac98-aea141142a36-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.141504 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/962f0d62-e23b-437e-ac98-aea141142a36-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.143314 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/962f0d62-e23b-437e-ac98-aea141142a36-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.157429 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/962f0d62-e23b-437e-ac98-aea141142a36-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.161779 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22n2c\" (UniqueName: \"kubernetes.io/projected/962f0d62-e23b-437e-ac98-aea141142a36-kube-api-access-22n2c\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.186406 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"962f0d62-e23b-437e-ac98-aea141142a36\") " pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.216853 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.629009 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27182112-178e-415c-9af9-f7fbf58d8120" path="/var/lib/kubelet/pods/27182112-178e-415c-9af9-f7fbf58d8120/volumes" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.631086 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f32d61be-992e-4a09-9663-33d3d59779e9" path="/var/lib/kubelet/pods/f32d61be-992e-4a09-9663-33d3d59779e9/volumes" Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.682452 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 02 02:00:17 crc kubenswrapper[4884]: W1202 02:00:17.754925 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod962f0d62_e23b_437e_ac98_aea141142a36.slice/crio-3b36dff179ceb3c06ffe4da89fe3908d16713aada872a2b2257ab0510bc0b7d5 WatchSource:0}: Error finding container 3b36dff179ceb3c06ffe4da89fe3908d16713aada872a2b2257ab0510bc0b7d5: Status 404 returned error can't find the container with id 3b36dff179ceb3c06ffe4da89fe3908d16713aada872a2b2257ab0510bc0b7d5 Dec 02 02:00:17 crc kubenswrapper[4884]: I1202 02:00:17.825553 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"962f0d62-e23b-437e-ac98-aea141142a36","Type":"ContainerStarted","Data":"3b36dff179ceb3c06ffe4da89fe3908d16713aada872a2b2257ab0510bc0b7d5"} Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.434558 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-dkgs4"] Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.436491 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.443516 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.447616 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-dkgs4"] Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.550177 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.550495 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-config\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.550700 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvktk\" (UniqueName: \"kubernetes.io/projected/87ad2c99-23ed-4bf0-8698-411180f67f3b-kube-api-access-cvktk\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.550880 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.550989 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.551278 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.551429 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.653232 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-config\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.653350 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvktk\" (UniqueName: \"kubernetes.io/projected/87ad2c99-23ed-4bf0-8698-411180f67f3b-kube-api-access-cvktk\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.653406 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.653423 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.653448 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.653487 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.653514 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.654341 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-openstack-edpm-ipam\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.654972 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-dns-svc\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.655091 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-ovsdbserver-nb\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.655296 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-config\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.655798 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-dns-swift-storage-0\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.656354 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-ovsdbserver-sb\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.680357 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvktk\" (UniqueName: \"kubernetes.io/projected/87ad2c99-23ed-4bf0-8698-411180f67f3b-kube-api-access-cvktk\") pod \"dnsmasq-dns-79bd4cc8c9-dkgs4\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.756533 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:18 crc kubenswrapper[4884]: I1202 02:00:18.835923 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c","Type":"ContainerStarted","Data":"cab059b2a8f449377d81a5a967892bc113a07d5d6a00a95e0c385dd02ae2ed42"} Dec 02 02:00:19 crc kubenswrapper[4884]: I1202 02:00:19.242588 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-dkgs4"] Dec 02 02:00:19 crc kubenswrapper[4884]: I1202 02:00:19.846257 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"962f0d62-e23b-437e-ac98-aea141142a36","Type":"ContainerStarted","Data":"c6a3e0b7ac5d26112a5a85ca4ac2e3426910f495416d4250553f41473ec7db69"} Dec 02 02:00:19 crc kubenswrapper[4884]: I1202 02:00:19.849412 4884 generic.go:334] "Generic (PLEG): container finished" podID="87ad2c99-23ed-4bf0-8698-411180f67f3b" containerID="0246a6caa0fff944019f9aef5acb89f6b26bf1630527c86b4c29d5320179b497" exitCode=0 Dec 02 02:00:19 crc kubenswrapper[4884]: I1202 02:00:19.850866 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" event={"ID":"87ad2c99-23ed-4bf0-8698-411180f67f3b","Type":"ContainerDied","Data":"0246a6caa0fff944019f9aef5acb89f6b26bf1630527c86b4c29d5320179b497"} Dec 02 02:00:19 crc kubenswrapper[4884]: I1202 02:00:19.850891 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" event={"ID":"87ad2c99-23ed-4bf0-8698-411180f67f3b","Type":"ContainerStarted","Data":"ed47f2c8ed46f839909dd1e0c1bf0f3aaa0e6f13e6ef4b8acd3b4037813a9141"} Dec 02 02:00:20 crc kubenswrapper[4884]: I1202 02:00:20.864609 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" event={"ID":"87ad2c99-23ed-4bf0-8698-411180f67f3b","Type":"ContainerStarted","Data":"370bec02a4c219eef5268bc71db099edac2849ee5d6253e34a70e75b1c57712f"} Dec 02 02:00:20 crc kubenswrapper[4884]: I1202 02:00:20.864917 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:20 crc kubenswrapper[4884]: I1202 02:00:20.895507 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" podStartSLOduration=2.89549012 podStartE2EDuration="2.89549012s" podCreationTimestamp="2025-12-02 02:00:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 02:00:20.889685688 +0000 UTC m=+1437.565522572" watchObservedRunningTime="2025-12-02 02:00:20.89549012 +0000 UTC m=+1437.571327014" Dec 02 02:00:28 crc kubenswrapper[4884]: I1202 02:00:28.758212 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:28 crc kubenswrapper[4884]: I1202 02:00:28.837771 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-j6jlj"] Dec 02 02:00:28 crc kubenswrapper[4884]: I1202 02:00:28.838048 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" podUID="3e49c865-898a-496f-80b5-32c3ed00558b" containerName="dnsmasq-dns" containerID="cri-o://ca6e36d901f79f15f193e08db6cec23b2ac6ad6aa717ed56fb27891f1f22b341" gracePeriod=10 Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.046928 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6cd9bffc9-zqncs"] Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.048653 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.060702 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cd9bffc9-zqncs"] Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.111777 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-dns-svc\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.112097 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-openstack-edpm-ipam\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.112175 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjcdc\" (UniqueName: \"kubernetes.io/projected/0bfdff8a-051f-4d73-94c8-9472af62a278-kube-api-access-sjcdc\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.112198 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-dns-swift-storage-0\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.112220 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-ovsdbserver-sb\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.112549 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-ovsdbserver-nb\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.112728 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-config\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.214343 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjcdc\" (UniqueName: \"kubernetes.io/projected/0bfdff8a-051f-4d73-94c8-9472af62a278-kube-api-access-sjcdc\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.214396 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-dns-swift-storage-0\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.214423 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-ovsdbserver-sb\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.214494 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-ovsdbserver-nb\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.214536 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-config\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.214566 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-dns-svc\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.214605 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-openstack-edpm-ipam\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.215523 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-openstack-edpm-ipam\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.215864 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-dns-swift-storage-0\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.216380 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-dns-svc\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.216469 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-config\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.216947 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-ovsdbserver-nb\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.216976 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0bfdff8a-051f-4d73-94c8-9472af62a278-ovsdbserver-sb\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.233965 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjcdc\" (UniqueName: \"kubernetes.io/projected/0bfdff8a-051f-4d73-94c8-9472af62a278-kube-api-access-sjcdc\") pod \"dnsmasq-dns-6cd9bffc9-zqncs\" (UID: \"0bfdff8a-051f-4d73-94c8-9472af62a278\") " pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.375638 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.478922 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.637324 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-ovsdbserver-sb\") pod \"3e49c865-898a-496f-80b5-32c3ed00558b\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.637769 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgjm9\" (UniqueName: \"kubernetes.io/projected/3e49c865-898a-496f-80b5-32c3ed00558b-kube-api-access-fgjm9\") pod \"3e49c865-898a-496f-80b5-32c3ed00558b\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.637797 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-ovsdbserver-nb\") pod \"3e49c865-898a-496f-80b5-32c3ed00558b\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.637882 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-dns-svc\") pod \"3e49c865-898a-496f-80b5-32c3ed00558b\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.637947 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-dns-swift-storage-0\") pod \"3e49c865-898a-496f-80b5-32c3ed00558b\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.638016 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-config\") pod \"3e49c865-898a-496f-80b5-32c3ed00558b\" (UID: \"3e49c865-898a-496f-80b5-32c3ed00558b\") " Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.676734 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e49c865-898a-496f-80b5-32c3ed00558b-kube-api-access-fgjm9" (OuterVolumeSpecName: "kube-api-access-fgjm9") pod "3e49c865-898a-496f-80b5-32c3ed00558b" (UID: "3e49c865-898a-496f-80b5-32c3ed00558b"). InnerVolumeSpecName "kube-api-access-fgjm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.725706 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3e49c865-898a-496f-80b5-32c3ed00558b" (UID: "3e49c865-898a-496f-80b5-32c3ed00558b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.730066 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3e49c865-898a-496f-80b5-32c3ed00558b" (UID: "3e49c865-898a-496f-80b5-32c3ed00558b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.740778 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-config" (OuterVolumeSpecName: "config") pod "3e49c865-898a-496f-80b5-32c3ed00558b" (UID: "3e49c865-898a-496f-80b5-32c3ed00558b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.740882 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3e49c865-898a-496f-80b5-32c3ed00558b" (UID: "3e49c865-898a-496f-80b5-32c3ed00558b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.741091 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgjm9\" (UniqueName: \"kubernetes.io/projected/3e49c865-898a-496f-80b5-32c3ed00558b-kube-api-access-fgjm9\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.741115 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.741124 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.741132 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-config\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.741141 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.758834 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3e49c865-898a-496f-80b5-32c3ed00558b" (UID: "3e49c865-898a-496f-80b5-32c3ed00558b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:29 crc kubenswrapper[4884]: I1202 02:00:29.843489 4884 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3e49c865-898a-496f-80b5-32c3ed00558b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:30 crc kubenswrapper[4884]: I1202 02:00:30.025418 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cd9bffc9-zqncs"] Dec 02 02:00:30 crc kubenswrapper[4884]: I1202 02:00:30.029822 4884 generic.go:334] "Generic (PLEG): container finished" podID="3e49c865-898a-496f-80b5-32c3ed00558b" containerID="ca6e36d901f79f15f193e08db6cec23b2ac6ad6aa717ed56fb27891f1f22b341" exitCode=0 Dec 02 02:00:30 crc kubenswrapper[4884]: I1202 02:00:30.029871 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" event={"ID":"3e49c865-898a-496f-80b5-32c3ed00558b","Type":"ContainerDied","Data":"ca6e36d901f79f15f193e08db6cec23b2ac6ad6aa717ed56fb27891f1f22b341"} Dec 02 02:00:30 crc kubenswrapper[4884]: I1202 02:00:30.029904 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" event={"ID":"3e49c865-898a-496f-80b5-32c3ed00558b","Type":"ContainerDied","Data":"3407bfc3b3f882b7feff4a4a15c77e0700bbffb70a44374e80fbaa3cd77cf341"} Dec 02 02:00:30 crc kubenswrapper[4884]: I1202 02:00:30.029925 4884 scope.go:117] "RemoveContainer" containerID="ca6e36d901f79f15f193e08db6cec23b2ac6ad6aa717ed56fb27891f1f22b341" Dec 02 02:00:30 crc kubenswrapper[4884]: I1202 02:00:30.030063 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-j6jlj" Dec 02 02:00:30 crc kubenswrapper[4884]: I1202 02:00:30.152199 4884 scope.go:117] "RemoveContainer" containerID="54973ed23c0576f50ec34ce1ce3bc9ea2757bd5f4b3e2678e8c8129033cb1f7f" Dec 02 02:00:30 crc kubenswrapper[4884]: I1202 02:00:30.177898 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-j6jlj"] Dec 02 02:00:30 crc kubenswrapper[4884]: I1202 02:00:30.189122 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-j6jlj"] Dec 02 02:00:30 crc kubenswrapper[4884]: I1202 02:00:30.197934 4884 scope.go:117] "RemoveContainer" containerID="ca6e36d901f79f15f193e08db6cec23b2ac6ad6aa717ed56fb27891f1f22b341" Dec 02 02:00:30 crc kubenswrapper[4884]: E1202 02:00:30.198327 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca6e36d901f79f15f193e08db6cec23b2ac6ad6aa717ed56fb27891f1f22b341\": container with ID starting with ca6e36d901f79f15f193e08db6cec23b2ac6ad6aa717ed56fb27891f1f22b341 not found: ID does not exist" containerID="ca6e36d901f79f15f193e08db6cec23b2ac6ad6aa717ed56fb27891f1f22b341" Dec 02 02:00:30 crc kubenswrapper[4884]: I1202 02:00:30.198355 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca6e36d901f79f15f193e08db6cec23b2ac6ad6aa717ed56fb27891f1f22b341"} err="failed to get container status \"ca6e36d901f79f15f193e08db6cec23b2ac6ad6aa717ed56fb27891f1f22b341\": rpc error: code = NotFound desc = could not find container \"ca6e36d901f79f15f193e08db6cec23b2ac6ad6aa717ed56fb27891f1f22b341\": container with ID starting with ca6e36d901f79f15f193e08db6cec23b2ac6ad6aa717ed56fb27891f1f22b341 not found: ID does not exist" Dec 02 02:00:30 crc kubenswrapper[4884]: I1202 02:00:30.198377 4884 scope.go:117] "RemoveContainer" containerID="54973ed23c0576f50ec34ce1ce3bc9ea2757bd5f4b3e2678e8c8129033cb1f7f" Dec 02 02:00:30 crc kubenswrapper[4884]: E1202 02:00:30.198539 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54973ed23c0576f50ec34ce1ce3bc9ea2757bd5f4b3e2678e8c8129033cb1f7f\": container with ID starting with 54973ed23c0576f50ec34ce1ce3bc9ea2757bd5f4b3e2678e8c8129033cb1f7f not found: ID does not exist" containerID="54973ed23c0576f50ec34ce1ce3bc9ea2757bd5f4b3e2678e8c8129033cb1f7f" Dec 02 02:00:30 crc kubenswrapper[4884]: I1202 02:00:30.198560 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54973ed23c0576f50ec34ce1ce3bc9ea2757bd5f4b3e2678e8c8129033cb1f7f"} err="failed to get container status \"54973ed23c0576f50ec34ce1ce3bc9ea2757bd5f4b3e2678e8c8129033cb1f7f\": rpc error: code = NotFound desc = could not find container \"54973ed23c0576f50ec34ce1ce3bc9ea2757bd5f4b3e2678e8c8129033cb1f7f\": container with ID starting with 54973ed23c0576f50ec34ce1ce3bc9ea2757bd5f4b3e2678e8c8129033cb1f7f not found: ID does not exist" Dec 02 02:00:31 crc kubenswrapper[4884]: I1202 02:00:31.041235 4884 generic.go:334] "Generic (PLEG): container finished" podID="0bfdff8a-051f-4d73-94c8-9472af62a278" containerID="4738742e02deb3277e245955db09ee3987dc29438893d226b0804ddafd4afe64" exitCode=0 Dec 02 02:00:31 crc kubenswrapper[4884]: I1202 02:00:31.041361 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" event={"ID":"0bfdff8a-051f-4d73-94c8-9472af62a278","Type":"ContainerDied","Data":"4738742e02deb3277e245955db09ee3987dc29438893d226b0804ddafd4afe64"} Dec 02 02:00:31 crc kubenswrapper[4884]: I1202 02:00:31.041583 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" event={"ID":"0bfdff8a-051f-4d73-94c8-9472af62a278","Type":"ContainerStarted","Data":"ed821e955aef33221f4e99e631f076d2cb87eda60a5064fe7ac18fbc2561d7a3"} Dec 02 02:00:31 crc kubenswrapper[4884]: I1202 02:00:31.661944 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e49c865-898a-496f-80b5-32c3ed00558b" path="/var/lib/kubelet/pods/3e49c865-898a-496f-80b5-32c3ed00558b/volumes" Dec 02 02:00:32 crc kubenswrapper[4884]: I1202 02:00:32.056404 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" event={"ID":"0bfdff8a-051f-4d73-94c8-9472af62a278","Type":"ContainerStarted","Data":"3ada160ca7c9e01ba1d08084122579ff5550defa963a316d79bbc03f80d38ff7"} Dec 02 02:00:32 crc kubenswrapper[4884]: I1202 02:00:32.056615 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:32 crc kubenswrapper[4884]: I1202 02:00:32.079725 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" podStartSLOduration=3.079697712 podStartE2EDuration="3.079697712s" podCreationTimestamp="2025-12-02 02:00:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 02:00:32.072896555 +0000 UTC m=+1448.748733469" watchObservedRunningTime="2025-12-02 02:00:32.079697712 +0000 UTC m=+1448.755534616" Dec 02 02:00:39 crc kubenswrapper[4884]: I1202 02:00:39.378084 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6cd9bffc9-zqncs" Dec 02 02:00:39 crc kubenswrapper[4884]: I1202 02:00:39.507015 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-dkgs4"] Dec 02 02:00:39 crc kubenswrapper[4884]: I1202 02:00:39.507301 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" podUID="87ad2c99-23ed-4bf0-8698-411180f67f3b" containerName="dnsmasq-dns" containerID="cri-o://370bec02a4c219eef5268bc71db099edac2849ee5d6253e34a70e75b1c57712f" gracePeriod=10 Dec 02 02:00:39 crc kubenswrapper[4884]: I1202 02:00:39.976628 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.061487 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-ovsdbserver-sb\") pod \"87ad2c99-23ed-4bf0-8698-411180f67f3b\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.061575 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-dns-swift-storage-0\") pod \"87ad2c99-23ed-4bf0-8698-411180f67f3b\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.061605 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-config\") pod \"87ad2c99-23ed-4bf0-8698-411180f67f3b\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.062915 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-dns-svc\") pod \"87ad2c99-23ed-4bf0-8698-411180f67f3b\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.062964 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-ovsdbserver-nb\") pod \"87ad2c99-23ed-4bf0-8698-411180f67f3b\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.063066 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-openstack-edpm-ipam\") pod \"87ad2c99-23ed-4bf0-8698-411180f67f3b\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.063115 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvktk\" (UniqueName: \"kubernetes.io/projected/87ad2c99-23ed-4bf0-8698-411180f67f3b-kube-api-access-cvktk\") pod \"87ad2c99-23ed-4bf0-8698-411180f67f3b\" (UID: \"87ad2c99-23ed-4bf0-8698-411180f67f3b\") " Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.077320 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87ad2c99-23ed-4bf0-8698-411180f67f3b-kube-api-access-cvktk" (OuterVolumeSpecName: "kube-api-access-cvktk") pod "87ad2c99-23ed-4bf0-8698-411180f67f3b" (UID: "87ad2c99-23ed-4bf0-8698-411180f67f3b"). InnerVolumeSpecName "kube-api-access-cvktk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.131222 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "87ad2c99-23ed-4bf0-8698-411180f67f3b" (UID: "87ad2c99-23ed-4bf0-8698-411180f67f3b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.138237 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "87ad2c99-23ed-4bf0-8698-411180f67f3b" (UID: "87ad2c99-23ed-4bf0-8698-411180f67f3b"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.141542 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "87ad2c99-23ed-4bf0-8698-411180f67f3b" (UID: "87ad2c99-23ed-4bf0-8698-411180f67f3b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.152037 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "87ad2c99-23ed-4bf0-8698-411180f67f3b" (UID: "87ad2c99-23ed-4bf0-8698-411180f67f3b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.156229 4884 generic.go:334] "Generic (PLEG): container finished" podID="87ad2c99-23ed-4bf0-8698-411180f67f3b" containerID="370bec02a4c219eef5268bc71db099edac2849ee5d6253e34a70e75b1c57712f" exitCode=0 Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.156287 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" event={"ID":"87ad2c99-23ed-4bf0-8698-411180f67f3b","Type":"ContainerDied","Data":"370bec02a4c219eef5268bc71db099edac2849ee5d6253e34a70e75b1c57712f"} Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.156369 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.156420 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79bd4cc8c9-dkgs4" event={"ID":"87ad2c99-23ed-4bf0-8698-411180f67f3b","Type":"ContainerDied","Data":"ed47f2c8ed46f839909dd1e0c1bf0f3aaa0e6f13e6ef4b8acd3b4037813a9141"} Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.156459 4884 scope.go:117] "RemoveContainer" containerID="370bec02a4c219eef5268bc71db099edac2849ee5d6253e34a70e75b1c57712f" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.164799 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "87ad2c99-23ed-4bf0-8698-411180f67f3b" (UID: "87ad2c99-23ed-4bf0-8698-411180f67f3b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.165072 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.165099 4884 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.165110 4884 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.165119 4884 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.165126 4884 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.165135 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvktk\" (UniqueName: \"kubernetes.io/projected/87ad2c99-23ed-4bf0-8698-411180f67f3b-kube-api-access-cvktk\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.174384 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-config" (OuterVolumeSpecName: "config") pod "87ad2c99-23ed-4bf0-8698-411180f67f3b" (UID: "87ad2c99-23ed-4bf0-8698-411180f67f3b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.208937 4884 scope.go:117] "RemoveContainer" containerID="0246a6caa0fff944019f9aef5acb89f6b26bf1630527c86b4c29d5320179b497" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.229058 4884 scope.go:117] "RemoveContainer" containerID="370bec02a4c219eef5268bc71db099edac2849ee5d6253e34a70e75b1c57712f" Dec 02 02:00:40 crc kubenswrapper[4884]: E1202 02:00:40.229576 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"370bec02a4c219eef5268bc71db099edac2849ee5d6253e34a70e75b1c57712f\": container with ID starting with 370bec02a4c219eef5268bc71db099edac2849ee5d6253e34a70e75b1c57712f not found: ID does not exist" containerID="370bec02a4c219eef5268bc71db099edac2849ee5d6253e34a70e75b1c57712f" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.229643 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"370bec02a4c219eef5268bc71db099edac2849ee5d6253e34a70e75b1c57712f"} err="failed to get container status \"370bec02a4c219eef5268bc71db099edac2849ee5d6253e34a70e75b1c57712f\": rpc error: code = NotFound desc = could not find container \"370bec02a4c219eef5268bc71db099edac2849ee5d6253e34a70e75b1c57712f\": container with ID starting with 370bec02a4c219eef5268bc71db099edac2849ee5d6253e34a70e75b1c57712f not found: ID does not exist" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.229685 4884 scope.go:117] "RemoveContainer" containerID="0246a6caa0fff944019f9aef5acb89f6b26bf1630527c86b4c29d5320179b497" Dec 02 02:00:40 crc kubenswrapper[4884]: E1202 02:00:40.230136 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0246a6caa0fff944019f9aef5acb89f6b26bf1630527c86b4c29d5320179b497\": container with ID starting with 0246a6caa0fff944019f9aef5acb89f6b26bf1630527c86b4c29d5320179b497 not found: ID does not exist" containerID="0246a6caa0fff944019f9aef5acb89f6b26bf1630527c86b4c29d5320179b497" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.230174 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0246a6caa0fff944019f9aef5acb89f6b26bf1630527c86b4c29d5320179b497"} err="failed to get container status \"0246a6caa0fff944019f9aef5acb89f6b26bf1630527c86b4c29d5320179b497\": rpc error: code = NotFound desc = could not find container \"0246a6caa0fff944019f9aef5acb89f6b26bf1630527c86b4c29d5320179b497\": container with ID starting with 0246a6caa0fff944019f9aef5acb89f6b26bf1630527c86b4c29d5320179b497 not found: ID does not exist" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.267371 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87ad2c99-23ed-4bf0-8698-411180f67f3b-config\") on node \"crc\" DevicePath \"\"" Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.509139 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-dkgs4"] Dec 02 02:00:40 crc kubenswrapper[4884]: I1202 02:00:40.518367 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79bd4cc8c9-dkgs4"] Dec 02 02:00:41 crc kubenswrapper[4884]: I1202 02:00:41.634835 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87ad2c99-23ed-4bf0-8698-411180f67f3b" path="/var/lib/kubelet/pods/87ad2c99-23ed-4bf0-8698-411180f67f3b/volumes" Dec 02 02:00:46 crc kubenswrapper[4884]: I1202 02:00:46.971718 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:00:46 crc kubenswrapper[4884]: I1202 02:00:46.972222 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:00:46 crc kubenswrapper[4884]: I1202 02:00:46.972272 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 02:00:46 crc kubenswrapper[4884]: I1202 02:00:46.973110 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5e633af905e3298525fa039bc54a435b8ffe0999ec52daa8e4c43c66ebc12f1d"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 02:00:46 crc kubenswrapper[4884]: I1202 02:00:46.973174 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://5e633af905e3298525fa039bc54a435b8ffe0999ec52daa8e4c43c66ebc12f1d" gracePeriod=600 Dec 02 02:00:47 crc kubenswrapper[4884]: I1202 02:00:47.270152 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="5e633af905e3298525fa039bc54a435b8ffe0999ec52daa8e4c43c66ebc12f1d" exitCode=0 Dec 02 02:00:47 crc kubenswrapper[4884]: I1202 02:00:47.270244 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"5e633af905e3298525fa039bc54a435b8ffe0999ec52daa8e4c43c66ebc12f1d"} Dec 02 02:00:47 crc kubenswrapper[4884]: I1202 02:00:47.270607 4884 scope.go:117] "RemoveContainer" containerID="58124d3cd233d6efb787d810e5402863819796c49f3516cfa49fd6248e566581" Dec 02 02:00:48 crc kubenswrapper[4884]: I1202 02:00:48.285480 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac"} Dec 02 02:00:51 crc kubenswrapper[4884]: I1202 02:00:51.318565 4884 generic.go:334] "Generic (PLEG): container finished" podID="8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c" containerID="cab059b2a8f449377d81a5a967892bc113a07d5d6a00a95e0c385dd02ae2ed42" exitCode=0 Dec 02 02:00:51 crc kubenswrapper[4884]: I1202 02:00:51.318645 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c","Type":"ContainerDied","Data":"cab059b2a8f449377d81a5a967892bc113a07d5d6a00a95e0c385dd02ae2ed42"} Dec 02 02:00:52 crc kubenswrapper[4884]: I1202 02:00:52.338432 4884 generic.go:334] "Generic (PLEG): container finished" podID="962f0d62-e23b-437e-ac98-aea141142a36" containerID="c6a3e0b7ac5d26112a5a85ca4ac2e3426910f495416d4250553f41473ec7db69" exitCode=0 Dec 02 02:00:52 crc kubenswrapper[4884]: I1202 02:00:52.339006 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"962f0d62-e23b-437e-ac98-aea141142a36","Type":"ContainerDied","Data":"c6a3e0b7ac5d26112a5a85ca4ac2e3426910f495416d4250553f41473ec7db69"} Dec 02 02:00:52 crc kubenswrapper[4884]: I1202 02:00:52.357442 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c","Type":"ContainerStarted","Data":"4b9713e1e74d391ed9c05e52e379af61a59c9faee4380ef88eea01e18a7522cc"} Dec 02 02:00:52 crc kubenswrapper[4884]: I1202 02:00:52.357807 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 02 02:00:52 crc kubenswrapper[4884]: I1202 02:00:52.417004 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.416985321 podStartE2EDuration="37.416985321s" podCreationTimestamp="2025-12-02 02:00:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 02:00:52.413307651 +0000 UTC m=+1469.089144565" watchObservedRunningTime="2025-12-02 02:00:52.416985321 +0000 UTC m=+1469.092822205" Dec 02 02:00:53 crc kubenswrapper[4884]: I1202 02:00:53.372638 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"962f0d62-e23b-437e-ac98-aea141142a36","Type":"ContainerStarted","Data":"8cefcbf36f1168c33c9147b70dd6697a906bd92c8a3346391352f07c56ae6a8e"} Dec 02 02:00:53 crc kubenswrapper[4884]: I1202 02:00:53.373363 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:00:53 crc kubenswrapper[4884]: I1202 02:00:53.414235 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.414211518 podStartE2EDuration="37.414211518s" podCreationTimestamp="2025-12-02 02:00:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 02:00:53.40125053 +0000 UTC m=+1470.077087484" watchObservedRunningTime="2025-12-02 02:00:53.414211518 +0000 UTC m=+1470.090048442" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.822182 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rgk7h"] Dec 02 02:00:54 crc kubenswrapper[4884]: E1202 02:00:54.823100 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e49c865-898a-496f-80b5-32c3ed00558b" containerName="dnsmasq-dns" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.823116 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e49c865-898a-496f-80b5-32c3ed00558b" containerName="dnsmasq-dns" Dec 02 02:00:54 crc kubenswrapper[4884]: E1202 02:00:54.823137 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87ad2c99-23ed-4bf0-8698-411180f67f3b" containerName="init" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.823145 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="87ad2c99-23ed-4bf0-8698-411180f67f3b" containerName="init" Dec 02 02:00:54 crc kubenswrapper[4884]: E1202 02:00:54.823183 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87ad2c99-23ed-4bf0-8698-411180f67f3b" containerName="dnsmasq-dns" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.823191 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="87ad2c99-23ed-4bf0-8698-411180f67f3b" containerName="dnsmasq-dns" Dec 02 02:00:54 crc kubenswrapper[4884]: E1202 02:00:54.823212 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e49c865-898a-496f-80b5-32c3ed00558b" containerName="init" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.823220 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e49c865-898a-496f-80b5-32c3ed00558b" containerName="init" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.823513 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="87ad2c99-23ed-4bf0-8698-411180f67f3b" containerName="dnsmasq-dns" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.823534 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e49c865-898a-496f-80b5-32c3ed00558b" containerName="dnsmasq-dns" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.825397 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.853001 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rgk7h"] Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.896857 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-utilities\") pod \"certified-operators-rgk7h\" (UID: \"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c\") " pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.896943 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skg8x\" (UniqueName: \"kubernetes.io/projected/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-kube-api-access-skg8x\") pod \"certified-operators-rgk7h\" (UID: \"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c\") " pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.897069 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-catalog-content\") pod \"certified-operators-rgk7h\" (UID: \"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c\") " pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.998627 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skg8x\" (UniqueName: \"kubernetes.io/projected/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-kube-api-access-skg8x\") pod \"certified-operators-rgk7h\" (UID: \"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c\") " pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.998801 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-catalog-content\") pod \"certified-operators-rgk7h\" (UID: \"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c\") " pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.998873 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-utilities\") pod \"certified-operators-rgk7h\" (UID: \"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c\") " pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.999316 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-catalog-content\") pod \"certified-operators-rgk7h\" (UID: \"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c\") " pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:00:54 crc kubenswrapper[4884]: I1202 02:00:54.999449 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-utilities\") pod \"certified-operators-rgk7h\" (UID: \"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c\") " pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:00:55 crc kubenswrapper[4884]: I1202 02:00:55.029717 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skg8x\" (UniqueName: \"kubernetes.io/projected/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-kube-api-access-skg8x\") pod \"certified-operators-rgk7h\" (UID: \"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c\") " pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:00:55 crc kubenswrapper[4884]: I1202 02:00:55.166762 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:00:55 crc kubenswrapper[4884]: I1202 02:00:55.719776 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rgk7h"] Dec 02 02:00:55 crc kubenswrapper[4884]: W1202 02:00:55.720077 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode83e5c5f_a5bd_4b52_88f3_60f7edfaf55c.slice/crio-30b19d17a4b67b0ba6106fab4831c1fda521eebe3b20a3d167526209c091c74e WatchSource:0}: Error finding container 30b19d17a4b67b0ba6106fab4831c1fda521eebe3b20a3d167526209c091c74e: Status 404 returned error can't find the container with id 30b19d17a4b67b0ba6106fab4831c1fda521eebe3b20a3d167526209c091c74e Dec 02 02:00:56 crc kubenswrapper[4884]: I1202 02:00:56.447435 4884 generic.go:334] "Generic (PLEG): container finished" podID="e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" containerID="54fbdd402c06f9c267102123fc75fa31ff90810d30307609423b52153c123f67" exitCode=0 Dec 02 02:00:56 crc kubenswrapper[4884]: I1202 02:00:56.447716 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgk7h" event={"ID":"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c","Type":"ContainerDied","Data":"54fbdd402c06f9c267102123fc75fa31ff90810d30307609423b52153c123f67"} Dec 02 02:00:56 crc kubenswrapper[4884]: I1202 02:00:56.448337 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgk7h" event={"ID":"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c","Type":"ContainerStarted","Data":"30b19d17a4b67b0ba6106fab4831c1fda521eebe3b20a3d167526209c091c74e"} Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.462356 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgk7h" event={"ID":"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c","Type":"ContainerStarted","Data":"7af177aaaea6921f6db2e07ce8cddcf75b34b242928f146919db6cd098f50104"} Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.680657 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62"] Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.682631 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.684282 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.684582 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.684704 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.687523 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.688275 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62"] Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.758795 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.758938 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dn2w\" (UniqueName: \"kubernetes.io/projected/06fb15ba-f86b-4962-8b6b-52367ad24e09-kube-api-access-6dn2w\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.759048 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.759088 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.860727 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.860901 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.860976 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dn2w\" (UniqueName: \"kubernetes.io/projected/06fb15ba-f86b-4962-8b6b-52367ad24e09-kube-api-access-6dn2w\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.861066 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.866305 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.866310 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.867162 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:00:57 crc kubenswrapper[4884]: I1202 02:00:57.884635 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dn2w\" (UniqueName: \"kubernetes.io/projected/06fb15ba-f86b-4962-8b6b-52367ad24e09-kube-api-access-6dn2w\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:00:58 crc kubenswrapper[4884]: I1202 02:00:58.000966 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:00:58 crc kubenswrapper[4884]: I1202 02:00:58.475181 4884 generic.go:334] "Generic (PLEG): container finished" podID="e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" containerID="7af177aaaea6921f6db2e07ce8cddcf75b34b242928f146919db6cd098f50104" exitCode=0 Dec 02 02:00:58 crc kubenswrapper[4884]: I1202 02:00:58.475243 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgk7h" event={"ID":"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c","Type":"ContainerDied","Data":"7af177aaaea6921f6db2e07ce8cddcf75b34b242928f146919db6cd098f50104"} Dec 02 02:00:58 crc kubenswrapper[4884]: I1202 02:00:58.723265 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62"] Dec 02 02:00:59 crc kubenswrapper[4884]: I1202 02:00:59.486227 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" event={"ID":"06fb15ba-f86b-4962-8b6b-52367ad24e09","Type":"ContainerStarted","Data":"0d1422088e7094100866354d0b101ffe035a0efc684dea7f2958cb957ee0bce7"} Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.148686 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29410681-v4hzm"] Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.150632 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.214054 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29410681-v4hzm"] Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.316166 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-combined-ca-bundle\") pod \"keystone-cron-29410681-v4hzm\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.316403 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2rrg\" (UniqueName: \"kubernetes.io/projected/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-kube-api-access-b2rrg\") pod \"keystone-cron-29410681-v4hzm\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.316505 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-config-data\") pod \"keystone-cron-29410681-v4hzm\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.316691 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-fernet-keys\") pod \"keystone-cron-29410681-v4hzm\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.418893 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-combined-ca-bundle\") pod \"keystone-cron-29410681-v4hzm\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.418964 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2rrg\" (UniqueName: \"kubernetes.io/projected/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-kube-api-access-b2rrg\") pod \"keystone-cron-29410681-v4hzm\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.418998 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-config-data\") pod \"keystone-cron-29410681-v4hzm\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.419081 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-fernet-keys\") pod \"keystone-cron-29410681-v4hzm\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.431802 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-combined-ca-bundle\") pod \"keystone-cron-29410681-v4hzm\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.431920 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-fernet-keys\") pod \"keystone-cron-29410681-v4hzm\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.436284 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2rrg\" (UniqueName: \"kubernetes.io/projected/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-kube-api-access-b2rrg\") pod \"keystone-cron-29410681-v4hzm\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.436298 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-config-data\") pod \"keystone-cron-29410681-v4hzm\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.472607 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:00 crc kubenswrapper[4884]: I1202 02:01:00.499535 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgk7h" event={"ID":"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c","Type":"ContainerStarted","Data":"af3c8dad38e46563279bef953c04a19f11a148cec3ffab1455463718710f2594"} Dec 02 02:01:01 crc kubenswrapper[4884]: I1202 02:01:01.018589 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rgk7h" podStartSLOduration=3.84427102 podStartE2EDuration="7.018556148s" podCreationTimestamp="2025-12-02 02:00:54 +0000 UTC" firstStartedPulling="2025-12-02 02:00:56.453261309 +0000 UTC m=+1473.129098223" lastFinishedPulling="2025-12-02 02:00:59.627546467 +0000 UTC m=+1476.303383351" observedRunningTime="2025-12-02 02:01:00.523127852 +0000 UTC m=+1477.198964736" watchObservedRunningTime="2025-12-02 02:01:01.018556148 +0000 UTC m=+1477.694393032" Dec 02 02:01:01 crc kubenswrapper[4884]: I1202 02:01:01.025553 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29410681-v4hzm"] Dec 02 02:01:01 crc kubenswrapper[4884]: I1202 02:01:01.800719 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jz5rw"] Dec 02 02:01:01 crc kubenswrapper[4884]: I1202 02:01:01.803510 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:01 crc kubenswrapper[4884]: I1202 02:01:01.810586 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jz5rw"] Dec 02 02:01:01 crc kubenswrapper[4884]: I1202 02:01:01.949122 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-catalog-content\") pod \"redhat-marketplace-jz5rw\" (UID: \"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c\") " pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:01 crc kubenswrapper[4884]: I1202 02:01:01.949266 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-utilities\") pod \"redhat-marketplace-jz5rw\" (UID: \"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c\") " pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:01 crc kubenswrapper[4884]: I1202 02:01:01.949294 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qbnr\" (UniqueName: \"kubernetes.io/projected/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-kube-api-access-9qbnr\") pod \"redhat-marketplace-jz5rw\" (UID: \"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c\") " pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:02 crc kubenswrapper[4884]: I1202 02:01:02.051432 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-utilities\") pod \"redhat-marketplace-jz5rw\" (UID: \"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c\") " pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:02 crc kubenswrapper[4884]: I1202 02:01:02.051487 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qbnr\" (UniqueName: \"kubernetes.io/projected/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-kube-api-access-9qbnr\") pod \"redhat-marketplace-jz5rw\" (UID: \"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c\") " pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:02 crc kubenswrapper[4884]: I1202 02:01:02.051599 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-catalog-content\") pod \"redhat-marketplace-jz5rw\" (UID: \"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c\") " pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:02 crc kubenswrapper[4884]: I1202 02:01:02.051881 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-utilities\") pod \"redhat-marketplace-jz5rw\" (UID: \"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c\") " pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:02 crc kubenswrapper[4884]: I1202 02:01:02.052326 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-catalog-content\") pod \"redhat-marketplace-jz5rw\" (UID: \"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c\") " pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:02 crc kubenswrapper[4884]: I1202 02:01:02.075930 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qbnr\" (UniqueName: \"kubernetes.io/projected/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-kube-api-access-9qbnr\") pod \"redhat-marketplace-jz5rw\" (UID: \"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c\") " pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:02 crc kubenswrapper[4884]: I1202 02:01:02.131587 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:02 crc kubenswrapper[4884]: W1202 02:01:02.181867 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fdf7bbf_bbc0_4280_99b9_8c87c8f6635c.slice/crio-3de0a3cc8e99b91e252a5edcf8935e93facb24e103028474dcbef939ffa6b5a7 WatchSource:0}: Error finding container 3de0a3cc8e99b91e252a5edcf8935e93facb24e103028474dcbef939ffa6b5a7: Status 404 returned error can't find the container with id 3de0a3cc8e99b91e252a5edcf8935e93facb24e103028474dcbef939ffa6b5a7 Dec 02 02:01:02 crc kubenswrapper[4884]: I1202 02:01:02.523219 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410681-v4hzm" event={"ID":"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c","Type":"ContainerStarted","Data":"3de0a3cc8e99b91e252a5edcf8935e93facb24e103028474dcbef939ffa6b5a7"} Dec 02 02:01:02 crc kubenswrapper[4884]: I1202 02:01:02.759083 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jz5rw"] Dec 02 02:01:03 crc kubenswrapper[4884]: I1202 02:01:03.537865 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410681-v4hzm" event={"ID":"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c","Type":"ContainerStarted","Data":"1ff74d279c2e3fb9a5b168e24b23a2e2081f29488c3bbf4ed10612d7cc59f1ea"} Dec 02 02:01:03 crc kubenswrapper[4884]: I1202 02:01:03.556541 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29410681-v4hzm" podStartSLOduration=3.556517335 podStartE2EDuration="3.556517335s" podCreationTimestamp="2025-12-02 02:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 02:01:03.556375632 +0000 UTC m=+1480.232212516" watchObservedRunningTime="2025-12-02 02:01:03.556517335 +0000 UTC m=+1480.232354229" Dec 02 02:01:05 crc kubenswrapper[4884]: I1202 02:01:05.167602 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:01:05 crc kubenswrapper[4884]: I1202 02:01:05.167937 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:01:05 crc kubenswrapper[4884]: I1202 02:01:05.564182 4884 generic.go:334] "Generic (PLEG): container finished" podID="8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c" containerID="1ff74d279c2e3fb9a5b168e24b23a2e2081f29488c3bbf4ed10612d7cc59f1ea" exitCode=0 Dec 02 02:01:05 crc kubenswrapper[4884]: I1202 02:01:05.564273 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410681-v4hzm" event={"ID":"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c","Type":"ContainerDied","Data":"1ff74d279c2e3fb9a5b168e24b23a2e2081f29488c3bbf4ed10612d7cc59f1ea"} Dec 02 02:01:06 crc kubenswrapper[4884]: I1202 02:01:06.230850 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-rgk7h" podUID="e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" containerName="registry-server" probeResult="failure" output=< Dec 02 02:01:06 crc kubenswrapper[4884]: timeout: failed to connect service ":50051" within 1s Dec 02 02:01:06 crc kubenswrapper[4884]: > Dec 02 02:01:06 crc kubenswrapper[4884]: I1202 02:01:06.236931 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 02 02:01:07 crc kubenswrapper[4884]: I1202 02:01:07.218966 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 02 02:01:09 crc kubenswrapper[4884]: I1202 02:01:09.645453 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410681-v4hzm" event={"ID":"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c","Type":"ContainerDied","Data":"3de0a3cc8e99b91e252a5edcf8935e93facb24e103028474dcbef939ffa6b5a7"} Dec 02 02:01:09 crc kubenswrapper[4884]: I1202 02:01:09.649598 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3de0a3cc8e99b91e252a5edcf8935e93facb24e103028474dcbef939ffa6b5a7" Dec 02 02:01:09 crc kubenswrapper[4884]: I1202 02:01:09.649629 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jz5rw" event={"ID":"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c","Type":"ContainerStarted","Data":"b4c005cfd9b34a923b10713a69570f4a22a90def69f6d709504d125de301f808"} Dec 02 02:01:09 crc kubenswrapper[4884]: I1202 02:01:09.917314 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:09 crc kubenswrapper[4884]: I1202 02:01:09.939130 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2rrg\" (UniqueName: \"kubernetes.io/projected/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-kube-api-access-b2rrg\") pod \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " Dec 02 02:01:09 crc kubenswrapper[4884]: I1202 02:01:09.939200 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-combined-ca-bundle\") pod \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " Dec 02 02:01:09 crc kubenswrapper[4884]: I1202 02:01:09.939277 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-fernet-keys\") pod \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " Dec 02 02:01:09 crc kubenswrapper[4884]: I1202 02:01:09.939490 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-config-data\") pod \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\" (UID: \"8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c\") " Dec 02 02:01:09 crc kubenswrapper[4884]: I1202 02:01:09.965473 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-kube-api-access-b2rrg" (OuterVolumeSpecName: "kube-api-access-b2rrg") pod "8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c" (UID: "8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c"). InnerVolumeSpecName "kube-api-access-b2rrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:01:09 crc kubenswrapper[4884]: I1202 02:01:09.966044 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c" (UID: "8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:01:10 crc kubenswrapper[4884]: I1202 02:01:10.039822 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c" (UID: "8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:01:10 crc kubenswrapper[4884]: I1202 02:01:10.042690 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2rrg\" (UniqueName: \"kubernetes.io/projected/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-kube-api-access-b2rrg\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:10 crc kubenswrapper[4884]: I1202 02:01:10.042730 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:10 crc kubenswrapper[4884]: I1202 02:01:10.042773 4884 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:10 crc kubenswrapper[4884]: I1202 02:01:10.058675 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-config-data" (OuterVolumeSpecName: "config-data") pod "8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c" (UID: "8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:01:10 crc kubenswrapper[4884]: I1202 02:01:10.144869 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:10 crc kubenswrapper[4884]: I1202 02:01:10.630667 4884 generic.go:334] "Generic (PLEG): container finished" podID="f2db21c8-46c3-4ffa-849a-bf1d17b7f69c" containerID="7d4085da430181db636e54457a90c24a32ae528e184236c43c6f8cf61e600337" exitCode=0 Dec 02 02:01:10 crc kubenswrapper[4884]: I1202 02:01:10.630731 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jz5rw" event={"ID":"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c","Type":"ContainerDied","Data":"7d4085da430181db636e54457a90c24a32ae528e184236c43c6f8cf61e600337"} Dec 02 02:01:10 crc kubenswrapper[4884]: I1202 02:01:10.634721 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410681-v4hzm" Dec 02 02:01:10 crc kubenswrapper[4884]: I1202 02:01:10.635192 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" event={"ID":"06fb15ba-f86b-4962-8b6b-52367ad24e09","Type":"ContainerStarted","Data":"d0b8d402a3f8ed10c3ab9c8f71b8ab99b2019e24a020d7754f13295ee5d22c43"} Dec 02 02:01:10 crc kubenswrapper[4884]: I1202 02:01:10.682393 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" podStartSLOduration=2.833791582 podStartE2EDuration="13.682375936s" podCreationTimestamp="2025-12-02 02:00:57 +0000 UTC" firstStartedPulling="2025-12-02 02:00:58.72663316 +0000 UTC m=+1475.402470034" lastFinishedPulling="2025-12-02 02:01:09.575217474 +0000 UTC m=+1486.251054388" observedRunningTime="2025-12-02 02:01:10.675483957 +0000 UTC m=+1487.351320841" watchObservedRunningTime="2025-12-02 02:01:10.682375936 +0000 UTC m=+1487.358212830" Dec 02 02:01:11 crc kubenswrapper[4884]: I1202 02:01:11.648389 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jz5rw" event={"ID":"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c","Type":"ContainerStarted","Data":"fabf14bda167f95be49c0ff8474e7b5f6d9bd336478bb30cb8e612c0cf1ead7b"} Dec 02 02:01:12 crc kubenswrapper[4884]: I1202 02:01:12.661489 4884 generic.go:334] "Generic (PLEG): container finished" podID="f2db21c8-46c3-4ffa-849a-bf1d17b7f69c" containerID="fabf14bda167f95be49c0ff8474e7b5f6d9bd336478bb30cb8e612c0cf1ead7b" exitCode=0 Dec 02 02:01:12 crc kubenswrapper[4884]: I1202 02:01:12.661788 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jz5rw" event={"ID":"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c","Type":"ContainerDied","Data":"fabf14bda167f95be49c0ff8474e7b5f6d9bd336478bb30cb8e612c0cf1ead7b"} Dec 02 02:01:14 crc kubenswrapper[4884]: I1202 02:01:14.682463 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jz5rw" event={"ID":"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c","Type":"ContainerStarted","Data":"631035b84916f9d0155c5bbdd49bf1ed9e9f8179949120701b25d72888de7c92"} Dec 02 02:01:14 crc kubenswrapper[4884]: I1202 02:01:14.712184 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jz5rw" podStartSLOduration=10.734492249 podStartE2EDuration="13.712156997s" podCreationTimestamp="2025-12-02 02:01:01 +0000 UTC" firstStartedPulling="2025-12-02 02:01:10.633064267 +0000 UTC m=+1487.308901161" lastFinishedPulling="2025-12-02 02:01:13.610729025 +0000 UTC m=+1490.286565909" observedRunningTime="2025-12-02 02:01:14.702572192 +0000 UTC m=+1491.378409106" watchObservedRunningTime="2025-12-02 02:01:14.712156997 +0000 UTC m=+1491.387993911" Dec 02 02:01:15 crc kubenswrapper[4884]: I1202 02:01:15.239567 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:01:15 crc kubenswrapper[4884]: I1202 02:01:15.298315 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:01:15 crc kubenswrapper[4884]: I1202 02:01:15.929956 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rgk7h"] Dec 02 02:01:16 crc kubenswrapper[4884]: I1202 02:01:16.711663 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rgk7h" podUID="e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" containerName="registry-server" containerID="cri-o://af3c8dad38e46563279bef953c04a19f11a148cec3ffab1455463718710f2594" gracePeriod=2 Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.174560 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.303183 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skg8x\" (UniqueName: \"kubernetes.io/projected/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-kube-api-access-skg8x\") pod \"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c\" (UID: \"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c\") " Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.303658 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-catalog-content\") pod \"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c\" (UID: \"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c\") " Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.303905 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-utilities\") pod \"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c\" (UID: \"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c\") " Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.304272 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-utilities" (OuterVolumeSpecName: "utilities") pod "e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" (UID: "e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.304702 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.309449 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-kube-api-access-skg8x" (OuterVolumeSpecName: "kube-api-access-skg8x") pod "e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" (UID: "e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c"). InnerVolumeSpecName "kube-api-access-skg8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.349569 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" (UID: "e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.406784 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skg8x\" (UniqueName: \"kubernetes.io/projected/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-kube-api-access-skg8x\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.406825 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.726284 4884 generic.go:334] "Generic (PLEG): container finished" podID="e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" containerID="af3c8dad38e46563279bef953c04a19f11a148cec3ffab1455463718710f2594" exitCode=0 Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.726350 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgk7h" event={"ID":"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c","Type":"ContainerDied","Data":"af3c8dad38e46563279bef953c04a19f11a148cec3ffab1455463718710f2594"} Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.726401 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgk7h" event={"ID":"e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c","Type":"ContainerDied","Data":"30b19d17a4b67b0ba6106fab4831c1fda521eebe3b20a3d167526209c091c74e"} Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.726430 4884 scope.go:117] "RemoveContainer" containerID="af3c8dad38e46563279bef953c04a19f11a148cec3ffab1455463718710f2594" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.726421 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgk7h" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.761074 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rgk7h"] Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.766321 4884 scope.go:117] "RemoveContainer" containerID="7af177aaaea6921f6db2e07ce8cddcf75b34b242928f146919db6cd098f50104" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.771598 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rgk7h"] Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.785160 4884 scope.go:117] "RemoveContainer" containerID="54fbdd402c06f9c267102123fc75fa31ff90810d30307609423b52153c123f67" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.840924 4884 scope.go:117] "RemoveContainer" containerID="af3c8dad38e46563279bef953c04a19f11a148cec3ffab1455463718710f2594" Dec 02 02:01:17 crc kubenswrapper[4884]: E1202 02:01:17.841432 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af3c8dad38e46563279bef953c04a19f11a148cec3ffab1455463718710f2594\": container with ID starting with af3c8dad38e46563279bef953c04a19f11a148cec3ffab1455463718710f2594 not found: ID does not exist" containerID="af3c8dad38e46563279bef953c04a19f11a148cec3ffab1455463718710f2594" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.841490 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af3c8dad38e46563279bef953c04a19f11a148cec3ffab1455463718710f2594"} err="failed to get container status \"af3c8dad38e46563279bef953c04a19f11a148cec3ffab1455463718710f2594\": rpc error: code = NotFound desc = could not find container \"af3c8dad38e46563279bef953c04a19f11a148cec3ffab1455463718710f2594\": container with ID starting with af3c8dad38e46563279bef953c04a19f11a148cec3ffab1455463718710f2594 not found: ID does not exist" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.841526 4884 scope.go:117] "RemoveContainer" containerID="7af177aaaea6921f6db2e07ce8cddcf75b34b242928f146919db6cd098f50104" Dec 02 02:01:17 crc kubenswrapper[4884]: E1202 02:01:17.842170 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7af177aaaea6921f6db2e07ce8cddcf75b34b242928f146919db6cd098f50104\": container with ID starting with 7af177aaaea6921f6db2e07ce8cddcf75b34b242928f146919db6cd098f50104 not found: ID does not exist" containerID="7af177aaaea6921f6db2e07ce8cddcf75b34b242928f146919db6cd098f50104" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.842216 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7af177aaaea6921f6db2e07ce8cddcf75b34b242928f146919db6cd098f50104"} err="failed to get container status \"7af177aaaea6921f6db2e07ce8cddcf75b34b242928f146919db6cd098f50104\": rpc error: code = NotFound desc = could not find container \"7af177aaaea6921f6db2e07ce8cddcf75b34b242928f146919db6cd098f50104\": container with ID starting with 7af177aaaea6921f6db2e07ce8cddcf75b34b242928f146919db6cd098f50104 not found: ID does not exist" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.842245 4884 scope.go:117] "RemoveContainer" containerID="54fbdd402c06f9c267102123fc75fa31ff90810d30307609423b52153c123f67" Dec 02 02:01:17 crc kubenswrapper[4884]: E1202 02:01:17.842659 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54fbdd402c06f9c267102123fc75fa31ff90810d30307609423b52153c123f67\": container with ID starting with 54fbdd402c06f9c267102123fc75fa31ff90810d30307609423b52153c123f67 not found: ID does not exist" containerID="54fbdd402c06f9c267102123fc75fa31ff90810d30307609423b52153c123f67" Dec 02 02:01:17 crc kubenswrapper[4884]: I1202 02:01:17.842732 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54fbdd402c06f9c267102123fc75fa31ff90810d30307609423b52153c123f67"} err="failed to get container status \"54fbdd402c06f9c267102123fc75fa31ff90810d30307609423b52153c123f67\": rpc error: code = NotFound desc = could not find container \"54fbdd402c06f9c267102123fc75fa31ff90810d30307609423b52153c123f67\": container with ID starting with 54fbdd402c06f9c267102123fc75fa31ff90810d30307609423b52153c123f67 not found: ID does not exist" Dec 02 02:01:19 crc kubenswrapper[4884]: I1202 02:01:19.652289 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" path="/var/lib/kubelet/pods/e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c/volumes" Dec 02 02:01:22 crc kubenswrapper[4884]: I1202 02:01:22.132290 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:22 crc kubenswrapper[4884]: I1202 02:01:22.132645 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:22 crc kubenswrapper[4884]: I1202 02:01:22.195056 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:22 crc kubenswrapper[4884]: I1202 02:01:22.845432 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:22 crc kubenswrapper[4884]: I1202 02:01:22.906126 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jz5rw"] Dec 02 02:01:23 crc kubenswrapper[4884]: I1202 02:01:23.803884 4884 generic.go:334] "Generic (PLEG): container finished" podID="06fb15ba-f86b-4962-8b6b-52367ad24e09" containerID="d0b8d402a3f8ed10c3ab9c8f71b8ab99b2019e24a020d7754f13295ee5d22c43" exitCode=0 Dec 02 02:01:23 crc kubenswrapper[4884]: I1202 02:01:23.803987 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" event={"ID":"06fb15ba-f86b-4962-8b6b-52367ad24e09","Type":"ContainerDied","Data":"d0b8d402a3f8ed10c3ab9c8f71b8ab99b2019e24a020d7754f13295ee5d22c43"} Dec 02 02:01:24 crc kubenswrapper[4884]: I1202 02:01:24.825159 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jz5rw" podUID="f2db21c8-46c3-4ffa-849a-bf1d17b7f69c" containerName="registry-server" containerID="cri-o://631035b84916f9d0155c5bbdd49bf1ed9e9f8179949120701b25d72888de7c92" gracePeriod=2 Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.363543 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.369451 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.393796 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-inventory\") pod \"06fb15ba-f86b-4962-8b6b-52367ad24e09\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.393854 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-ssh-key\") pod \"06fb15ba-f86b-4962-8b6b-52367ad24e09\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.393884 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-utilities\") pod \"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c\" (UID: \"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c\") " Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.395879 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-utilities" (OuterVolumeSpecName: "utilities") pod "f2db21c8-46c3-4ffa-849a-bf1d17b7f69c" (UID: "f2db21c8-46c3-4ffa-849a-bf1d17b7f69c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.396863 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-catalog-content\") pod \"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c\" (UID: \"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c\") " Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.396930 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qbnr\" (UniqueName: \"kubernetes.io/projected/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-kube-api-access-9qbnr\") pod \"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c\" (UID: \"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c\") " Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.396979 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dn2w\" (UniqueName: \"kubernetes.io/projected/06fb15ba-f86b-4962-8b6b-52367ad24e09-kube-api-access-6dn2w\") pod \"06fb15ba-f86b-4962-8b6b-52367ad24e09\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.397035 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-repo-setup-combined-ca-bundle\") pod \"06fb15ba-f86b-4962-8b6b-52367ad24e09\" (UID: \"06fb15ba-f86b-4962-8b6b-52367ad24e09\") " Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.397627 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.409599 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "06fb15ba-f86b-4962-8b6b-52367ad24e09" (UID: "06fb15ba-f86b-4962-8b6b-52367ad24e09"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.413392 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-kube-api-access-9qbnr" (OuterVolumeSpecName: "kube-api-access-9qbnr") pod "f2db21c8-46c3-4ffa-849a-bf1d17b7f69c" (UID: "f2db21c8-46c3-4ffa-849a-bf1d17b7f69c"). InnerVolumeSpecName "kube-api-access-9qbnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.422083 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06fb15ba-f86b-4962-8b6b-52367ad24e09-kube-api-access-6dn2w" (OuterVolumeSpecName: "kube-api-access-6dn2w") pod "06fb15ba-f86b-4962-8b6b-52367ad24e09" (UID: "06fb15ba-f86b-4962-8b6b-52367ad24e09"). InnerVolumeSpecName "kube-api-access-6dn2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.435502 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f2db21c8-46c3-4ffa-849a-bf1d17b7f69c" (UID: "f2db21c8-46c3-4ffa-849a-bf1d17b7f69c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.438430 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-inventory" (OuterVolumeSpecName: "inventory") pod "06fb15ba-f86b-4962-8b6b-52367ad24e09" (UID: "06fb15ba-f86b-4962-8b6b-52367ad24e09"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.442923 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "06fb15ba-f86b-4962-8b6b-52367ad24e09" (UID: "06fb15ba-f86b-4962-8b6b-52367ad24e09"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.499830 4884 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.499865 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.499875 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06fb15ba-f86b-4962-8b6b-52367ad24e09-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.499884 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.499895 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qbnr\" (UniqueName: \"kubernetes.io/projected/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c-kube-api-access-9qbnr\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.499912 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dn2w\" (UniqueName: \"kubernetes.io/projected/06fb15ba-f86b-4962-8b6b-52367ad24e09-kube-api-access-6dn2w\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.847294 4884 generic.go:334] "Generic (PLEG): container finished" podID="f2db21c8-46c3-4ffa-849a-bf1d17b7f69c" containerID="631035b84916f9d0155c5bbdd49bf1ed9e9f8179949120701b25d72888de7c92" exitCode=0 Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.847533 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jz5rw" event={"ID":"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c","Type":"ContainerDied","Data":"631035b84916f9d0155c5bbdd49bf1ed9e9f8179949120701b25d72888de7c92"} Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.847961 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jz5rw" event={"ID":"f2db21c8-46c3-4ffa-849a-bf1d17b7f69c","Type":"ContainerDied","Data":"b4c005cfd9b34a923b10713a69570f4a22a90def69f6d709504d125de301f808"} Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.848042 4884 scope.go:117] "RemoveContainer" containerID="631035b84916f9d0155c5bbdd49bf1ed9e9f8179949120701b25d72888de7c92" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.847646 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jz5rw" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.855738 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" event={"ID":"06fb15ba-f86b-4962-8b6b-52367ad24e09","Type":"ContainerDied","Data":"0d1422088e7094100866354d0b101ffe035a0efc684dea7f2958cb957ee0bce7"} Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.855823 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d1422088e7094100866354d0b101ffe035a0efc684dea7f2958cb957ee0bce7" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.855954 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.899183 4884 scope.go:117] "RemoveContainer" containerID="fabf14bda167f95be49c0ff8474e7b5f6d9bd336478bb30cb8e612c0cf1ead7b" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.929299 4884 scope.go:117] "RemoveContainer" containerID="7d4085da430181db636e54457a90c24a32ae528e184236c43c6f8cf61e600337" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.940135 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jz5rw"] Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.955021 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jz5rw"] Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.963221 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv"] Dec 02 02:01:25 crc kubenswrapper[4884]: E1202 02:01:25.963716 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2db21c8-46c3-4ffa-849a-bf1d17b7f69c" containerName="extract-utilities" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.963737 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2db21c8-46c3-4ffa-849a-bf1d17b7f69c" containerName="extract-utilities" Dec 02 02:01:25 crc kubenswrapper[4884]: E1202 02:01:25.963773 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c" containerName="keystone-cron" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.963783 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c" containerName="keystone-cron" Dec 02 02:01:25 crc kubenswrapper[4884]: E1202 02:01:25.963801 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2db21c8-46c3-4ffa-849a-bf1d17b7f69c" containerName="registry-server" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.963809 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2db21c8-46c3-4ffa-849a-bf1d17b7f69c" containerName="registry-server" Dec 02 02:01:25 crc kubenswrapper[4884]: E1202 02:01:25.963829 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" containerName="registry-server" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.963837 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" containerName="registry-server" Dec 02 02:01:25 crc kubenswrapper[4884]: E1202 02:01:25.963857 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" containerName="extract-utilities" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.963865 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" containerName="extract-utilities" Dec 02 02:01:25 crc kubenswrapper[4884]: E1202 02:01:25.963888 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2db21c8-46c3-4ffa-849a-bf1d17b7f69c" containerName="extract-content" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.963897 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2db21c8-46c3-4ffa-849a-bf1d17b7f69c" containerName="extract-content" Dec 02 02:01:25 crc kubenswrapper[4884]: E1202 02:01:25.963915 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" containerName="extract-content" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.963923 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" containerName="extract-content" Dec 02 02:01:25 crc kubenswrapper[4884]: E1202 02:01:25.963938 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06fb15ba-f86b-4962-8b6b-52367ad24e09" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.963949 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="06fb15ba-f86b-4962-8b6b-52367ad24e09" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.964173 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="06fb15ba-f86b-4962-8b6b-52367ad24e09" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.964195 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="e83e5c5f-a5bd-4b52-88f3-60f7edfaf55c" containerName="registry-server" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.964214 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c" containerName="keystone-cron" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.964230 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2db21c8-46c3-4ffa-849a-bf1d17b7f69c" containerName="registry-server" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.965012 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.971973 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv"] Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.991995 4884 scope.go:117] "RemoveContainer" containerID="631035b84916f9d0155c5bbdd49bf1ed9e9f8179949120701b25d72888de7c92" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.992328 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.992395 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:01:25 crc kubenswrapper[4884]: E1202 02:01:25.992412 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"631035b84916f9d0155c5bbdd49bf1ed9e9f8179949120701b25d72888de7c92\": container with ID starting with 631035b84916f9d0155c5bbdd49bf1ed9e9f8179949120701b25d72888de7c92 not found: ID does not exist" containerID="631035b84916f9d0155c5bbdd49bf1ed9e9f8179949120701b25d72888de7c92" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.992439 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"631035b84916f9d0155c5bbdd49bf1ed9e9f8179949120701b25d72888de7c92"} err="failed to get container status \"631035b84916f9d0155c5bbdd49bf1ed9e9f8179949120701b25d72888de7c92\": rpc error: code = NotFound desc = could not find container \"631035b84916f9d0155c5bbdd49bf1ed9e9f8179949120701b25d72888de7c92\": container with ID starting with 631035b84916f9d0155c5bbdd49bf1ed9e9f8179949120701b25d72888de7c92 not found: ID does not exist" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.992459 4884 scope.go:117] "RemoveContainer" containerID="fabf14bda167f95be49c0ff8474e7b5f6d9bd336478bb30cb8e612c0cf1ead7b" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.992523 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.992606 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:01:25 crc kubenswrapper[4884]: E1202 02:01:25.993368 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fabf14bda167f95be49c0ff8474e7b5f6d9bd336478bb30cb8e612c0cf1ead7b\": container with ID starting with fabf14bda167f95be49c0ff8474e7b5f6d9bd336478bb30cb8e612c0cf1ead7b not found: ID does not exist" containerID="fabf14bda167f95be49c0ff8474e7b5f6d9bd336478bb30cb8e612c0cf1ead7b" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.993418 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fabf14bda167f95be49c0ff8474e7b5f6d9bd336478bb30cb8e612c0cf1ead7b"} err="failed to get container status \"fabf14bda167f95be49c0ff8474e7b5f6d9bd336478bb30cb8e612c0cf1ead7b\": rpc error: code = NotFound desc = could not find container \"fabf14bda167f95be49c0ff8474e7b5f6d9bd336478bb30cb8e612c0cf1ead7b\": container with ID starting with fabf14bda167f95be49c0ff8474e7b5f6d9bd336478bb30cb8e612c0cf1ead7b not found: ID does not exist" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.993462 4884 scope.go:117] "RemoveContainer" containerID="7d4085da430181db636e54457a90c24a32ae528e184236c43c6f8cf61e600337" Dec 02 02:01:25 crc kubenswrapper[4884]: E1202 02:01:25.994022 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d4085da430181db636e54457a90c24a32ae528e184236c43c6f8cf61e600337\": container with ID starting with 7d4085da430181db636e54457a90c24a32ae528e184236c43c6f8cf61e600337 not found: ID does not exist" containerID="7d4085da430181db636e54457a90c24a32ae528e184236c43c6f8cf61e600337" Dec 02 02:01:25 crc kubenswrapper[4884]: I1202 02:01:25.994063 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d4085da430181db636e54457a90c24a32ae528e184236c43c6f8cf61e600337"} err="failed to get container status \"7d4085da430181db636e54457a90c24a32ae528e184236c43c6f8cf61e600337\": rpc error: code = NotFound desc = could not find container \"7d4085da430181db636e54457a90c24a32ae528e184236c43c6f8cf61e600337\": container with ID starting with 7d4085da430181db636e54457a90c24a32ae528e184236c43c6f8cf61e600337 not found: ID does not exist" Dec 02 02:01:26 crc kubenswrapper[4884]: I1202 02:01:26.112518 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5rsj\" (UniqueName: \"kubernetes.io/projected/b074036a-792d-4270-af51-d3ae2dd92bda-kube-api-access-n5rsj\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ql9sv\" (UID: \"b074036a-792d-4270-af51-d3ae2dd92bda\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" Dec 02 02:01:26 crc kubenswrapper[4884]: I1202 02:01:26.112992 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b074036a-792d-4270-af51-d3ae2dd92bda-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ql9sv\" (UID: \"b074036a-792d-4270-af51-d3ae2dd92bda\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" Dec 02 02:01:26 crc kubenswrapper[4884]: I1202 02:01:26.113202 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b074036a-792d-4270-af51-d3ae2dd92bda-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ql9sv\" (UID: \"b074036a-792d-4270-af51-d3ae2dd92bda\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" Dec 02 02:01:26 crc kubenswrapper[4884]: I1202 02:01:26.216299 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b074036a-792d-4270-af51-d3ae2dd92bda-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ql9sv\" (UID: \"b074036a-792d-4270-af51-d3ae2dd92bda\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" Dec 02 02:01:26 crc kubenswrapper[4884]: I1202 02:01:26.216538 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b074036a-792d-4270-af51-d3ae2dd92bda-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ql9sv\" (UID: \"b074036a-792d-4270-af51-d3ae2dd92bda\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" Dec 02 02:01:26 crc kubenswrapper[4884]: I1202 02:01:26.216779 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5rsj\" (UniqueName: \"kubernetes.io/projected/b074036a-792d-4270-af51-d3ae2dd92bda-kube-api-access-n5rsj\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ql9sv\" (UID: \"b074036a-792d-4270-af51-d3ae2dd92bda\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" Dec 02 02:01:26 crc kubenswrapper[4884]: I1202 02:01:26.222440 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b074036a-792d-4270-af51-d3ae2dd92bda-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ql9sv\" (UID: \"b074036a-792d-4270-af51-d3ae2dd92bda\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" Dec 02 02:01:26 crc kubenswrapper[4884]: I1202 02:01:26.231676 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b074036a-792d-4270-af51-d3ae2dd92bda-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ql9sv\" (UID: \"b074036a-792d-4270-af51-d3ae2dd92bda\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" Dec 02 02:01:26 crc kubenswrapper[4884]: I1202 02:01:26.235138 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5rsj\" (UniqueName: \"kubernetes.io/projected/b074036a-792d-4270-af51-d3ae2dd92bda-kube-api-access-n5rsj\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-ql9sv\" (UID: \"b074036a-792d-4270-af51-d3ae2dd92bda\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" Dec 02 02:01:26 crc kubenswrapper[4884]: I1202 02:01:26.389013 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" Dec 02 02:01:27 crc kubenswrapper[4884]: W1202 02:01:27.053861 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb074036a_792d_4270_af51_d3ae2dd92bda.slice/crio-94e41cce624d0da667045292b96e3cd5c94d21e02a7e7af0dfe402748d9b0d45 WatchSource:0}: Error finding container 94e41cce624d0da667045292b96e3cd5c94d21e02a7e7af0dfe402748d9b0d45: Status 404 returned error can't find the container with id 94e41cce624d0da667045292b96e3cd5c94d21e02a7e7af0dfe402748d9b0d45 Dec 02 02:01:27 crc kubenswrapper[4884]: I1202 02:01:27.058056 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv"] Dec 02 02:01:27 crc kubenswrapper[4884]: I1202 02:01:27.632443 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2db21c8-46c3-4ffa-849a-bf1d17b7f69c" path="/var/lib/kubelet/pods/f2db21c8-46c3-4ffa-849a-bf1d17b7f69c/volumes" Dec 02 02:01:27 crc kubenswrapper[4884]: I1202 02:01:27.882144 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" event={"ID":"b074036a-792d-4270-af51-d3ae2dd92bda","Type":"ContainerStarted","Data":"a1843bf8c4db9f90b0135edeea359c85a4792c717ec0fe63e72b92cbf3bb18bc"} Dec 02 02:01:27 crc kubenswrapper[4884]: I1202 02:01:27.882192 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" event={"ID":"b074036a-792d-4270-af51-d3ae2dd92bda","Type":"ContainerStarted","Data":"94e41cce624d0da667045292b96e3cd5c94d21e02a7e7af0dfe402748d9b0d45"} Dec 02 02:01:27 crc kubenswrapper[4884]: I1202 02:01:27.912501 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" podStartSLOduration=2.41753432 podStartE2EDuration="2.912478773s" podCreationTimestamp="2025-12-02 02:01:25 +0000 UTC" firstStartedPulling="2025-12-02 02:01:27.057161045 +0000 UTC m=+1503.732997929" lastFinishedPulling="2025-12-02 02:01:27.552105498 +0000 UTC m=+1504.227942382" observedRunningTime="2025-12-02 02:01:27.904644491 +0000 UTC m=+1504.580481405" watchObservedRunningTime="2025-12-02 02:01:27.912478773 +0000 UTC m=+1504.588315667" Dec 02 02:01:30 crc kubenswrapper[4884]: I1202 02:01:30.917351 4884 generic.go:334] "Generic (PLEG): container finished" podID="b074036a-792d-4270-af51-d3ae2dd92bda" containerID="a1843bf8c4db9f90b0135edeea359c85a4792c717ec0fe63e72b92cbf3bb18bc" exitCode=0 Dec 02 02:01:30 crc kubenswrapper[4884]: I1202 02:01:30.917463 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" event={"ID":"b074036a-792d-4270-af51-d3ae2dd92bda","Type":"ContainerDied","Data":"a1843bf8c4db9f90b0135edeea359c85a4792c717ec0fe63e72b92cbf3bb18bc"} Dec 02 02:01:32 crc kubenswrapper[4884]: I1202 02:01:32.407840 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" Dec 02 02:01:32 crc kubenswrapper[4884]: I1202 02:01:32.480455 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5rsj\" (UniqueName: \"kubernetes.io/projected/b074036a-792d-4270-af51-d3ae2dd92bda-kube-api-access-n5rsj\") pod \"b074036a-792d-4270-af51-d3ae2dd92bda\" (UID: \"b074036a-792d-4270-af51-d3ae2dd92bda\") " Dec 02 02:01:32 crc kubenswrapper[4884]: I1202 02:01:32.480774 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b074036a-792d-4270-af51-d3ae2dd92bda-inventory\") pod \"b074036a-792d-4270-af51-d3ae2dd92bda\" (UID: \"b074036a-792d-4270-af51-d3ae2dd92bda\") " Dec 02 02:01:32 crc kubenswrapper[4884]: I1202 02:01:32.480860 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b074036a-792d-4270-af51-d3ae2dd92bda-ssh-key\") pod \"b074036a-792d-4270-af51-d3ae2dd92bda\" (UID: \"b074036a-792d-4270-af51-d3ae2dd92bda\") " Dec 02 02:01:32 crc kubenswrapper[4884]: I1202 02:01:32.501887 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b074036a-792d-4270-af51-d3ae2dd92bda-kube-api-access-n5rsj" (OuterVolumeSpecName: "kube-api-access-n5rsj") pod "b074036a-792d-4270-af51-d3ae2dd92bda" (UID: "b074036a-792d-4270-af51-d3ae2dd92bda"). InnerVolumeSpecName "kube-api-access-n5rsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:01:32 crc kubenswrapper[4884]: I1202 02:01:32.550356 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b074036a-792d-4270-af51-d3ae2dd92bda-inventory" (OuterVolumeSpecName: "inventory") pod "b074036a-792d-4270-af51-d3ae2dd92bda" (UID: "b074036a-792d-4270-af51-d3ae2dd92bda"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:01:32 crc kubenswrapper[4884]: I1202 02:01:32.561001 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b074036a-792d-4270-af51-d3ae2dd92bda-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b074036a-792d-4270-af51-d3ae2dd92bda" (UID: "b074036a-792d-4270-af51-d3ae2dd92bda"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:01:32 crc kubenswrapper[4884]: I1202 02:01:32.583987 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b074036a-792d-4270-af51-d3ae2dd92bda-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:32 crc kubenswrapper[4884]: I1202 02:01:32.584030 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5rsj\" (UniqueName: \"kubernetes.io/projected/b074036a-792d-4270-af51-d3ae2dd92bda-kube-api-access-n5rsj\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:32 crc kubenswrapper[4884]: I1202 02:01:32.584045 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b074036a-792d-4270-af51-d3ae2dd92bda-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:01:32 crc kubenswrapper[4884]: I1202 02:01:32.949130 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" event={"ID":"b074036a-792d-4270-af51-d3ae2dd92bda","Type":"ContainerDied","Data":"94e41cce624d0da667045292b96e3cd5c94d21e02a7e7af0dfe402748d9b0d45"} Dec 02 02:01:32 crc kubenswrapper[4884]: I1202 02:01:32.949490 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94e41cce624d0da667045292b96e3cd5c94d21e02a7e7af0dfe402748d9b0d45" Dec 02 02:01:32 crc kubenswrapper[4884]: I1202 02:01:32.949272 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-ql9sv" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.041413 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p"] Dec 02 02:01:33 crc kubenswrapper[4884]: E1202 02:01:33.041840 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b074036a-792d-4270-af51-d3ae2dd92bda" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.041860 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="b074036a-792d-4270-af51-d3ae2dd92bda" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.042097 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="b074036a-792d-4270-af51-d3ae2dd92bda" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.042771 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.044883 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.046806 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.047503 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.050726 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.079831 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p"] Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.196945 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wvq5\" (UniqueName: \"kubernetes.io/projected/49a3a2e3-b9ff-49bb-b850-5a516270f73b-kube-api-access-6wvq5\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.197019 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.197564 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.197636 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.298776 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.298836 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wvq5\" (UniqueName: \"kubernetes.io/projected/49a3a2e3-b9ff-49bb-b850-5a516270f73b-kube-api-access-6wvq5\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.298864 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.298962 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.304069 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.304914 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.305481 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.329399 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wvq5\" (UniqueName: \"kubernetes.io/projected/49a3a2e3-b9ff-49bb-b850-5a516270f73b-kube-api-access-6wvq5\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.386776 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.750222 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p"] Dec 02 02:01:33 crc kubenswrapper[4884]: I1202 02:01:33.962340 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" event={"ID":"49a3a2e3-b9ff-49bb-b850-5a516270f73b","Type":"ContainerStarted","Data":"4d0bbf635e647bfbfeb89e001806487438d3f5ecfba1d0839f720d020a478703"} Dec 02 02:01:34 crc kubenswrapper[4884]: I1202 02:01:34.972378 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" event={"ID":"49a3a2e3-b9ff-49bb-b850-5a516270f73b","Type":"ContainerStarted","Data":"2cd97105f2827de7385c5e83ffdcf79bf346a6b42452ecb376cef0b253cff61f"} Dec 02 02:01:34 crc kubenswrapper[4884]: I1202 02:01:34.992983 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" podStartSLOduration=1.299809998 podStartE2EDuration="1.992939809s" podCreationTimestamp="2025-12-02 02:01:33 +0000 UTC" firstStartedPulling="2025-12-02 02:01:33.74385401 +0000 UTC m=+1510.419690894" lastFinishedPulling="2025-12-02 02:01:34.436983821 +0000 UTC m=+1511.112820705" observedRunningTime="2025-12-02 02:01:34.984460831 +0000 UTC m=+1511.660297715" watchObservedRunningTime="2025-12-02 02:01:34.992939809 +0000 UTC m=+1511.668776693" Dec 02 02:01:41 crc kubenswrapper[4884]: I1202 02:01:41.754299 4884 scope.go:117] "RemoveContainer" containerID="2c260c8245a4227ce0886eb4629815934d6639a7790618f122cc64e375a82684" Dec 02 02:01:56 crc kubenswrapper[4884]: I1202 02:01:56.388909 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vgg5x"] Dec 02 02:01:56 crc kubenswrapper[4884]: I1202 02:01:56.392462 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:01:56 crc kubenswrapper[4884]: I1202 02:01:56.400615 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vgg5x"] Dec 02 02:01:56 crc kubenswrapper[4884]: I1202 02:01:56.526037 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-utilities\") pod \"community-operators-vgg5x\" (UID: \"828d71ae-3541-47f2-ba9c-e7eec0d2f8af\") " pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:01:56 crc kubenswrapper[4884]: I1202 02:01:56.526223 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-catalog-content\") pod \"community-operators-vgg5x\" (UID: \"828d71ae-3541-47f2-ba9c-e7eec0d2f8af\") " pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:01:56 crc kubenswrapper[4884]: I1202 02:01:56.526450 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z49sf\" (UniqueName: \"kubernetes.io/projected/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-kube-api-access-z49sf\") pod \"community-operators-vgg5x\" (UID: \"828d71ae-3541-47f2-ba9c-e7eec0d2f8af\") " pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:01:56 crc kubenswrapper[4884]: I1202 02:01:56.628117 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-utilities\") pod \"community-operators-vgg5x\" (UID: \"828d71ae-3541-47f2-ba9c-e7eec0d2f8af\") " pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:01:56 crc kubenswrapper[4884]: I1202 02:01:56.628193 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-catalog-content\") pod \"community-operators-vgg5x\" (UID: \"828d71ae-3541-47f2-ba9c-e7eec0d2f8af\") " pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:01:56 crc kubenswrapper[4884]: I1202 02:01:56.628257 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z49sf\" (UniqueName: \"kubernetes.io/projected/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-kube-api-access-z49sf\") pod \"community-operators-vgg5x\" (UID: \"828d71ae-3541-47f2-ba9c-e7eec0d2f8af\") " pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:01:56 crc kubenswrapper[4884]: I1202 02:01:56.628651 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-catalog-content\") pod \"community-operators-vgg5x\" (UID: \"828d71ae-3541-47f2-ba9c-e7eec0d2f8af\") " pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:01:56 crc kubenswrapper[4884]: I1202 02:01:56.628719 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-utilities\") pod \"community-operators-vgg5x\" (UID: \"828d71ae-3541-47f2-ba9c-e7eec0d2f8af\") " pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:01:56 crc kubenswrapper[4884]: I1202 02:01:56.647312 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z49sf\" (UniqueName: \"kubernetes.io/projected/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-kube-api-access-z49sf\") pod \"community-operators-vgg5x\" (UID: \"828d71ae-3541-47f2-ba9c-e7eec0d2f8af\") " pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:01:56 crc kubenswrapper[4884]: I1202 02:01:56.720897 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:01:57 crc kubenswrapper[4884]: I1202 02:01:57.200406 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vgg5x"] Dec 02 02:01:57 crc kubenswrapper[4884]: I1202 02:01:57.263006 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgg5x" event={"ID":"828d71ae-3541-47f2-ba9c-e7eec0d2f8af","Type":"ContainerStarted","Data":"a994b3c5f908c162491f51a1b183b3c30d2ad7ce636d0bcc290bd2dfd8285b8c"} Dec 02 02:01:58 crc kubenswrapper[4884]: I1202 02:01:58.275874 4884 generic.go:334] "Generic (PLEG): container finished" podID="828d71ae-3541-47f2-ba9c-e7eec0d2f8af" containerID="7ddbebbdb226e62312100c405d81eabe6810292fb5bf0d39de502e7118cfcfe3" exitCode=0 Dec 02 02:01:58 crc kubenswrapper[4884]: I1202 02:01:58.276060 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgg5x" event={"ID":"828d71ae-3541-47f2-ba9c-e7eec0d2f8af","Type":"ContainerDied","Data":"7ddbebbdb226e62312100c405d81eabe6810292fb5bf0d39de502e7118cfcfe3"} Dec 02 02:01:59 crc kubenswrapper[4884]: I1202 02:01:59.287245 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgg5x" event={"ID":"828d71ae-3541-47f2-ba9c-e7eec0d2f8af","Type":"ContainerStarted","Data":"25f2372ee51a4c5768b669be4d4d174b30afa531299a781e4442120f532eca26"} Dec 02 02:02:00 crc kubenswrapper[4884]: I1202 02:02:00.303312 4884 generic.go:334] "Generic (PLEG): container finished" podID="828d71ae-3541-47f2-ba9c-e7eec0d2f8af" containerID="25f2372ee51a4c5768b669be4d4d174b30afa531299a781e4442120f532eca26" exitCode=0 Dec 02 02:02:00 crc kubenswrapper[4884]: I1202 02:02:00.303540 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgg5x" event={"ID":"828d71ae-3541-47f2-ba9c-e7eec0d2f8af","Type":"ContainerDied","Data":"25f2372ee51a4c5768b669be4d4d174b30afa531299a781e4442120f532eca26"} Dec 02 02:02:01 crc kubenswrapper[4884]: I1202 02:02:01.326342 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgg5x" event={"ID":"828d71ae-3541-47f2-ba9c-e7eec0d2f8af","Type":"ContainerStarted","Data":"775f56ad678f3f186540a0435ab2d96d54e436bc5bcc3e5998effc1cc50f9fe7"} Dec 02 02:02:01 crc kubenswrapper[4884]: I1202 02:02:01.356896 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vgg5x" podStartSLOduration=2.764064147 podStartE2EDuration="5.35687546s" podCreationTimestamp="2025-12-02 02:01:56 +0000 UTC" firstStartedPulling="2025-12-02 02:01:58.278040092 +0000 UTC m=+1534.953876976" lastFinishedPulling="2025-12-02 02:02:00.870851395 +0000 UTC m=+1537.546688289" observedRunningTime="2025-12-02 02:02:01.352491693 +0000 UTC m=+1538.028328617" watchObservedRunningTime="2025-12-02 02:02:01.35687546 +0000 UTC m=+1538.032712344" Dec 02 02:02:06 crc kubenswrapper[4884]: I1202 02:02:06.721428 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:02:06 crc kubenswrapper[4884]: I1202 02:02:06.721969 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:02:06 crc kubenswrapper[4884]: I1202 02:02:06.779041 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:02:07 crc kubenswrapper[4884]: I1202 02:02:07.467531 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:02:07 crc kubenswrapper[4884]: I1202 02:02:07.555270 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vgg5x"] Dec 02 02:02:09 crc kubenswrapper[4884]: I1202 02:02:09.408698 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vgg5x" podUID="828d71ae-3541-47f2-ba9c-e7eec0d2f8af" containerName="registry-server" containerID="cri-o://775f56ad678f3f186540a0435ab2d96d54e436bc5bcc3e5998effc1cc50f9fe7" gracePeriod=2 Dec 02 02:02:09 crc kubenswrapper[4884]: I1202 02:02:09.930628 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.007487 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-utilities\") pod \"828d71ae-3541-47f2-ba9c-e7eec0d2f8af\" (UID: \"828d71ae-3541-47f2-ba9c-e7eec0d2f8af\") " Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.007625 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z49sf\" (UniqueName: \"kubernetes.io/projected/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-kube-api-access-z49sf\") pod \"828d71ae-3541-47f2-ba9c-e7eec0d2f8af\" (UID: \"828d71ae-3541-47f2-ba9c-e7eec0d2f8af\") " Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.008011 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-catalog-content\") pod \"828d71ae-3541-47f2-ba9c-e7eec0d2f8af\" (UID: \"828d71ae-3541-47f2-ba9c-e7eec0d2f8af\") " Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.008424 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-utilities" (OuterVolumeSpecName: "utilities") pod "828d71ae-3541-47f2-ba9c-e7eec0d2f8af" (UID: "828d71ae-3541-47f2-ba9c-e7eec0d2f8af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.008844 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.014076 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-kube-api-access-z49sf" (OuterVolumeSpecName: "kube-api-access-z49sf") pod "828d71ae-3541-47f2-ba9c-e7eec0d2f8af" (UID: "828d71ae-3541-47f2-ba9c-e7eec0d2f8af"). InnerVolumeSpecName "kube-api-access-z49sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.074069 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "828d71ae-3541-47f2-ba9c-e7eec0d2f8af" (UID: "828d71ae-3541-47f2-ba9c-e7eec0d2f8af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.110234 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z49sf\" (UniqueName: \"kubernetes.io/projected/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-kube-api-access-z49sf\") on node \"crc\" DevicePath \"\"" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.110261 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/828d71ae-3541-47f2-ba9c-e7eec0d2f8af-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.418012 4884 generic.go:334] "Generic (PLEG): container finished" podID="828d71ae-3541-47f2-ba9c-e7eec0d2f8af" containerID="775f56ad678f3f186540a0435ab2d96d54e436bc5bcc3e5998effc1cc50f9fe7" exitCode=0 Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.418049 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgg5x" event={"ID":"828d71ae-3541-47f2-ba9c-e7eec0d2f8af","Type":"ContainerDied","Data":"775f56ad678f3f186540a0435ab2d96d54e436bc5bcc3e5998effc1cc50f9fe7"} Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.418096 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgg5x" event={"ID":"828d71ae-3541-47f2-ba9c-e7eec0d2f8af","Type":"ContainerDied","Data":"a994b3c5f908c162491f51a1b183b3c30d2ad7ce636d0bcc290bd2dfd8285b8c"} Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.418107 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgg5x" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.418114 4884 scope.go:117] "RemoveContainer" containerID="775f56ad678f3f186540a0435ab2d96d54e436bc5bcc3e5998effc1cc50f9fe7" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.436395 4884 scope.go:117] "RemoveContainer" containerID="25f2372ee51a4c5768b669be4d4d174b30afa531299a781e4442120f532eca26" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.465692 4884 scope.go:117] "RemoveContainer" containerID="7ddbebbdb226e62312100c405d81eabe6810292fb5bf0d39de502e7118cfcfe3" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.471568 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vgg5x"] Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.484450 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vgg5x"] Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.519197 4884 scope.go:117] "RemoveContainer" containerID="775f56ad678f3f186540a0435ab2d96d54e436bc5bcc3e5998effc1cc50f9fe7" Dec 02 02:02:10 crc kubenswrapper[4884]: E1202 02:02:10.519642 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"775f56ad678f3f186540a0435ab2d96d54e436bc5bcc3e5998effc1cc50f9fe7\": container with ID starting with 775f56ad678f3f186540a0435ab2d96d54e436bc5bcc3e5998effc1cc50f9fe7 not found: ID does not exist" containerID="775f56ad678f3f186540a0435ab2d96d54e436bc5bcc3e5998effc1cc50f9fe7" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.519690 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"775f56ad678f3f186540a0435ab2d96d54e436bc5bcc3e5998effc1cc50f9fe7"} err="failed to get container status \"775f56ad678f3f186540a0435ab2d96d54e436bc5bcc3e5998effc1cc50f9fe7\": rpc error: code = NotFound desc = could not find container \"775f56ad678f3f186540a0435ab2d96d54e436bc5bcc3e5998effc1cc50f9fe7\": container with ID starting with 775f56ad678f3f186540a0435ab2d96d54e436bc5bcc3e5998effc1cc50f9fe7 not found: ID does not exist" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.519722 4884 scope.go:117] "RemoveContainer" containerID="25f2372ee51a4c5768b669be4d4d174b30afa531299a781e4442120f532eca26" Dec 02 02:02:10 crc kubenswrapper[4884]: E1202 02:02:10.520124 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25f2372ee51a4c5768b669be4d4d174b30afa531299a781e4442120f532eca26\": container with ID starting with 25f2372ee51a4c5768b669be4d4d174b30afa531299a781e4442120f532eca26 not found: ID does not exist" containerID="25f2372ee51a4c5768b669be4d4d174b30afa531299a781e4442120f532eca26" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.520153 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25f2372ee51a4c5768b669be4d4d174b30afa531299a781e4442120f532eca26"} err="failed to get container status \"25f2372ee51a4c5768b669be4d4d174b30afa531299a781e4442120f532eca26\": rpc error: code = NotFound desc = could not find container \"25f2372ee51a4c5768b669be4d4d174b30afa531299a781e4442120f532eca26\": container with ID starting with 25f2372ee51a4c5768b669be4d4d174b30afa531299a781e4442120f532eca26 not found: ID does not exist" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.520170 4884 scope.go:117] "RemoveContainer" containerID="7ddbebbdb226e62312100c405d81eabe6810292fb5bf0d39de502e7118cfcfe3" Dec 02 02:02:10 crc kubenswrapper[4884]: E1202 02:02:10.520511 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ddbebbdb226e62312100c405d81eabe6810292fb5bf0d39de502e7118cfcfe3\": container with ID starting with 7ddbebbdb226e62312100c405d81eabe6810292fb5bf0d39de502e7118cfcfe3 not found: ID does not exist" containerID="7ddbebbdb226e62312100c405d81eabe6810292fb5bf0d39de502e7118cfcfe3" Dec 02 02:02:10 crc kubenswrapper[4884]: I1202 02:02:10.520546 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ddbebbdb226e62312100c405d81eabe6810292fb5bf0d39de502e7118cfcfe3"} err="failed to get container status \"7ddbebbdb226e62312100c405d81eabe6810292fb5bf0d39de502e7118cfcfe3\": rpc error: code = NotFound desc = could not find container \"7ddbebbdb226e62312100c405d81eabe6810292fb5bf0d39de502e7118cfcfe3\": container with ID starting with 7ddbebbdb226e62312100c405d81eabe6810292fb5bf0d39de502e7118cfcfe3 not found: ID does not exist" Dec 02 02:02:11 crc kubenswrapper[4884]: I1202 02:02:11.634703 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="828d71ae-3541-47f2-ba9c-e7eec0d2f8af" path="/var/lib/kubelet/pods/828d71ae-3541-47f2-ba9c-e7eec0d2f8af/volumes" Dec 02 02:02:41 crc kubenswrapper[4884]: I1202 02:02:41.918283 4884 scope.go:117] "RemoveContainer" containerID="c3a7e3020f57948047ca21a9940c2734138a0269a3009feb28ee61e353d08eb8" Dec 02 02:03:16 crc kubenswrapper[4884]: I1202 02:03:16.971821 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:03:16 crc kubenswrapper[4884]: I1202 02:03:16.972546 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:03:42 crc kubenswrapper[4884]: I1202 02:03:42.529155 4884 scope.go:117] "RemoveContainer" containerID="fb94676af85e01bb962d107a2d9516e111f83b880482244f79a8c359fcf6ee8a" Dec 02 02:03:46 crc kubenswrapper[4884]: I1202 02:03:46.972353 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:03:46 crc kubenswrapper[4884]: I1202 02:03:46.972968 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:04:16 crc kubenswrapper[4884]: I1202 02:04:16.972169 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:04:16 crc kubenswrapper[4884]: I1202 02:04:16.972885 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:04:16 crc kubenswrapper[4884]: I1202 02:04:16.972950 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 02:04:16 crc kubenswrapper[4884]: I1202 02:04:16.974003 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 02:04:16 crc kubenswrapper[4884]: I1202 02:04:16.974104 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" gracePeriod=600 Dec 02 02:04:17 crc kubenswrapper[4884]: E1202 02:04:17.119280 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:04:18 crc kubenswrapper[4884]: I1202 02:04:18.008142 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" exitCode=0 Dec 02 02:04:18 crc kubenswrapper[4884]: I1202 02:04:18.008217 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac"} Dec 02 02:04:18 crc kubenswrapper[4884]: I1202 02:04:18.008478 4884 scope.go:117] "RemoveContainer" containerID="5e633af905e3298525fa039bc54a435b8ffe0999ec52daa8e4c43c66ebc12f1d" Dec 02 02:04:18 crc kubenswrapper[4884]: I1202 02:04:18.009135 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:04:18 crc kubenswrapper[4884]: E1202 02:04:18.009367 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:04:30 crc kubenswrapper[4884]: I1202 02:04:30.615017 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:04:30 crc kubenswrapper[4884]: E1202 02:04:30.616359 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:04:41 crc kubenswrapper[4884]: I1202 02:04:41.615208 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:04:41 crc kubenswrapper[4884]: E1202 02:04:41.620895 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:04:42 crc kubenswrapper[4884]: I1202 02:04:42.618234 4884 scope.go:117] "RemoveContainer" containerID="bf92762a1db3ee254be4c737489b66d5ec6b31d47378a862216842b18ac20a7a" Dec 02 02:04:54 crc kubenswrapper[4884]: I1202 02:04:54.614612 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:04:54 crc kubenswrapper[4884]: E1202 02:04:54.615847 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:05:09 crc kubenswrapper[4884]: I1202 02:05:09.614242 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:05:09 crc kubenswrapper[4884]: E1202 02:05:09.615213 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:05:23 crc kubenswrapper[4884]: I1202 02:05:23.621002 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:05:23 crc kubenswrapper[4884]: E1202 02:05:23.621694 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.086641 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-8g74n"] Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.103351 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-dvqcv"] Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.113082 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-8g74n"] Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.122844 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-4d9f-account-create-update-bc8h8"] Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.134561 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-create-484hp"] Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.144605 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-4c10-account-create-update-kmph9"] Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.153209 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-4d9f-account-create-update-bc8h8"] Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.162979 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-dvqcv"] Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.173073 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-9g2b5"] Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.181484 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-355c-account-create-update-d5b2t"] Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.189274 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8791-account-create-update-j27h4"] Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.196653 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-create-484hp"] Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.203830 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-4c10-account-create-update-kmph9"] Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.213242 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-9g2b5"] Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.222151 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8791-account-create-update-j27h4"] Dec 02 02:05:28 crc kubenswrapper[4884]: I1202 02:05:28.248678 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-355c-account-create-update-d5b2t"] Dec 02 02:05:29 crc kubenswrapper[4884]: I1202 02:05:29.627126 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12142209-d074-426f-8e5d-d9e4acae21a0" path="/var/lib/kubelet/pods/12142209-d074-426f-8e5d-d9e4acae21a0/volumes" Dec 02 02:05:29 crc kubenswrapper[4884]: I1202 02:05:29.628074 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c11795b-a2e2-4906-b3fd-3048ec29a764" path="/var/lib/kubelet/pods/1c11795b-a2e2-4906-b3fd-3048ec29a764/volumes" Dec 02 02:05:29 crc kubenswrapper[4884]: I1202 02:05:29.628620 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33e4ef6f-578b-4149-8d23-fcf1b3f4a137" path="/var/lib/kubelet/pods/33e4ef6f-578b-4149-8d23-fcf1b3f4a137/volumes" Dec 02 02:05:29 crc kubenswrapper[4884]: I1202 02:05:29.629195 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f9006f4-9196-4a39-8c4d-10c4344a3b1f" path="/var/lib/kubelet/pods/7f9006f4-9196-4a39-8c4d-10c4344a3b1f/volumes" Dec 02 02:05:29 crc kubenswrapper[4884]: I1202 02:05:29.630537 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9922aaf4-6776-4e6c-ac62-82c43bef4697" path="/var/lib/kubelet/pods/9922aaf4-6776-4e6c-ac62-82c43bef4697/volumes" Dec 02 02:05:29 crc kubenswrapper[4884]: I1202 02:05:29.631360 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0bb5aee-14d4-4023-bb79-fdd1824aae0e" path="/var/lib/kubelet/pods/a0bb5aee-14d4-4023-bb79-fdd1824aae0e/volumes" Dec 02 02:05:29 crc kubenswrapper[4884]: I1202 02:05:29.632249 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c764da5f-6cfc-48aa-8a10-bf6bd97a226f" path="/var/lib/kubelet/pods/c764da5f-6cfc-48aa-8a10-bf6bd97a226f/volumes" Dec 02 02:05:29 crc kubenswrapper[4884]: I1202 02:05:29.633662 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05" path="/var/lib/kubelet/pods/e2ed98c4-d4b2-4fd3-8d0e-a6ed376bcf05/volumes" Dec 02 02:05:34 crc kubenswrapper[4884]: I1202 02:05:34.615037 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:05:34 crc kubenswrapper[4884]: E1202 02:05:34.616266 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:05:41 crc kubenswrapper[4884]: I1202 02:05:41.962224 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vj2gg"] Dec 02 02:05:41 crc kubenswrapper[4884]: E1202 02:05:41.963603 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="828d71ae-3541-47f2-ba9c-e7eec0d2f8af" containerName="extract-content" Dec 02 02:05:41 crc kubenswrapper[4884]: I1202 02:05:41.963627 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="828d71ae-3541-47f2-ba9c-e7eec0d2f8af" containerName="extract-content" Dec 02 02:05:41 crc kubenswrapper[4884]: E1202 02:05:41.963665 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="828d71ae-3541-47f2-ba9c-e7eec0d2f8af" containerName="extract-utilities" Dec 02 02:05:41 crc kubenswrapper[4884]: I1202 02:05:41.963678 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="828d71ae-3541-47f2-ba9c-e7eec0d2f8af" containerName="extract-utilities" Dec 02 02:05:41 crc kubenswrapper[4884]: E1202 02:05:41.963701 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="828d71ae-3541-47f2-ba9c-e7eec0d2f8af" containerName="registry-server" Dec 02 02:05:41 crc kubenswrapper[4884]: I1202 02:05:41.963715 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="828d71ae-3541-47f2-ba9c-e7eec0d2f8af" containerName="registry-server" Dec 02 02:05:41 crc kubenswrapper[4884]: I1202 02:05:41.964262 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="828d71ae-3541-47f2-ba9c-e7eec0d2f8af" containerName="registry-server" Dec 02 02:05:41 crc kubenswrapper[4884]: I1202 02:05:41.977477 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:05:41 crc kubenswrapper[4884]: I1202 02:05:41.977918 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vj2gg"] Dec 02 02:05:42 crc kubenswrapper[4884]: I1202 02:05:42.102328 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-catalog-content\") pod \"redhat-operators-vj2gg\" (UID: \"f6a2c03c-5d9e-43ec-8b75-63e536b06de5\") " pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:05:42 crc kubenswrapper[4884]: I1202 02:05:42.102621 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-utilities\") pod \"redhat-operators-vj2gg\" (UID: \"f6a2c03c-5d9e-43ec-8b75-63e536b06de5\") " pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:05:42 crc kubenswrapper[4884]: I1202 02:05:42.102885 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j4nw\" (UniqueName: \"kubernetes.io/projected/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-kube-api-access-6j4nw\") pod \"redhat-operators-vj2gg\" (UID: \"f6a2c03c-5d9e-43ec-8b75-63e536b06de5\") " pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:05:42 crc kubenswrapper[4884]: I1202 02:05:42.205545 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j4nw\" (UniqueName: \"kubernetes.io/projected/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-kube-api-access-6j4nw\") pod \"redhat-operators-vj2gg\" (UID: \"f6a2c03c-5d9e-43ec-8b75-63e536b06de5\") " pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:05:42 crc kubenswrapper[4884]: I1202 02:05:42.206476 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-catalog-content\") pod \"redhat-operators-vj2gg\" (UID: \"f6a2c03c-5d9e-43ec-8b75-63e536b06de5\") " pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:05:42 crc kubenswrapper[4884]: I1202 02:05:42.206584 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-utilities\") pod \"redhat-operators-vj2gg\" (UID: \"f6a2c03c-5d9e-43ec-8b75-63e536b06de5\") " pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:05:42 crc kubenswrapper[4884]: I1202 02:05:42.207294 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-catalog-content\") pod \"redhat-operators-vj2gg\" (UID: \"f6a2c03c-5d9e-43ec-8b75-63e536b06de5\") " pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:05:42 crc kubenswrapper[4884]: I1202 02:05:42.207343 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-utilities\") pod \"redhat-operators-vj2gg\" (UID: \"f6a2c03c-5d9e-43ec-8b75-63e536b06de5\") " pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:05:42 crc kubenswrapper[4884]: I1202 02:05:42.240556 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j4nw\" (UniqueName: \"kubernetes.io/projected/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-kube-api-access-6j4nw\") pod \"redhat-operators-vj2gg\" (UID: \"f6a2c03c-5d9e-43ec-8b75-63e536b06de5\") " pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:05:42 crc kubenswrapper[4884]: I1202 02:05:42.309710 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:05:42 crc kubenswrapper[4884]: I1202 02:05:42.705180 4884 scope.go:117] "RemoveContainer" containerID="27ee94a09469808f8d08fa4e19c1a06b2a625b5badff4d05ceb225571b42d96e" Dec 02 02:05:42 crc kubenswrapper[4884]: I1202 02:05:42.731435 4884 scope.go:117] "RemoveContainer" containerID="0009a1795b6b89769018f7bf4a7a02b1bb07a26d36f296e87ebb240ca8cc84bf" Dec 02 02:05:43 crc kubenswrapper[4884]: I1202 02:05:43.292972 4884 scope.go:117] "RemoveContainer" containerID="73dd6369a1ef29bdb1afed9fc8e62c11012312a273172268c8b4cf1b454abd4b" Dec 02 02:05:43 crc kubenswrapper[4884]: I1202 02:05:43.314050 4884 scope.go:117] "RemoveContainer" containerID="eea4c4a5937e70803834d77ed39c5a35944c2c5183e5e7b9c57c75d5072f961f" Dec 02 02:05:43 crc kubenswrapper[4884]: I1202 02:05:43.343987 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vj2gg"] Dec 02 02:05:43 crc kubenswrapper[4884]: I1202 02:05:43.428655 4884 scope.go:117] "RemoveContainer" containerID="7a4aa8b576a5faba0676a0107a1e59151a893f698b06b25a73e03328a444cc50" Dec 02 02:05:43 crc kubenswrapper[4884]: I1202 02:05:43.508365 4884 scope.go:117] "RemoveContainer" containerID="253198cc432448f0046c67fd4e99c31cf153588dc26bfb91548aa7651d20489c" Dec 02 02:05:43 crc kubenswrapper[4884]: I1202 02:05:43.586316 4884 scope.go:117] "RemoveContainer" containerID="99b6d45eb80119a04d35704de720f35dc87bad121a63e34001944bb5d691f4a0" Dec 02 02:05:43 crc kubenswrapper[4884]: I1202 02:05:43.621816 4884 scope.go:117] "RemoveContainer" containerID="d483bb9d56e217aa939f6edd2769b9496f0c550ec29df470b9b4461c520bfc7b" Dec 02 02:05:44 crc kubenswrapper[4884]: I1202 02:05:44.097728 4884 generic.go:334] "Generic (PLEG): container finished" podID="f6a2c03c-5d9e-43ec-8b75-63e536b06de5" containerID="786f216469c9d3e2ef5647cffb0a17791c5402a76953bbfa944cec47dde899cc" exitCode=0 Dec 02 02:05:44 crc kubenswrapper[4884]: I1202 02:05:44.098090 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj2gg" event={"ID":"f6a2c03c-5d9e-43ec-8b75-63e536b06de5","Type":"ContainerDied","Data":"786f216469c9d3e2ef5647cffb0a17791c5402a76953bbfa944cec47dde899cc"} Dec 02 02:05:44 crc kubenswrapper[4884]: I1202 02:05:44.098225 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj2gg" event={"ID":"f6a2c03c-5d9e-43ec-8b75-63e536b06de5","Type":"ContainerStarted","Data":"e190576b67d07a5869af47e97cd2b2c1a1a78f9315f7fa2df6874d9090643d95"} Dec 02 02:05:44 crc kubenswrapper[4884]: I1202 02:05:44.100344 4884 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 02:05:46 crc kubenswrapper[4884]: I1202 02:05:46.120300 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj2gg" event={"ID":"f6a2c03c-5d9e-43ec-8b75-63e536b06de5","Type":"ContainerStarted","Data":"516ece4494d6076b3c5031c61f71329de801b11c76c8d2b567e2706d2325d06f"} Dec 02 02:05:46 crc kubenswrapper[4884]: I1202 02:05:46.616509 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:05:46 crc kubenswrapper[4884]: E1202 02:05:46.617515 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:05:47 crc kubenswrapper[4884]: I1202 02:05:47.135640 4884 generic.go:334] "Generic (PLEG): container finished" podID="f6a2c03c-5d9e-43ec-8b75-63e536b06de5" containerID="516ece4494d6076b3c5031c61f71329de801b11c76c8d2b567e2706d2325d06f" exitCode=0 Dec 02 02:05:47 crc kubenswrapper[4884]: I1202 02:05:47.135735 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj2gg" event={"ID":"f6a2c03c-5d9e-43ec-8b75-63e536b06de5","Type":"ContainerDied","Data":"516ece4494d6076b3c5031c61f71329de801b11c76c8d2b567e2706d2325d06f"} Dec 02 02:05:49 crc kubenswrapper[4884]: I1202 02:05:49.161151 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj2gg" event={"ID":"f6a2c03c-5d9e-43ec-8b75-63e536b06de5","Type":"ContainerStarted","Data":"d941e7beb770c93595314ac8ca0962904699f434a234d88ad6117322fc685905"} Dec 02 02:05:49 crc kubenswrapper[4884]: I1202 02:05:49.188664 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vj2gg" podStartSLOduration=3.703868707 podStartE2EDuration="8.188645416s" podCreationTimestamp="2025-12-02 02:05:41 +0000 UTC" firstStartedPulling="2025-12-02 02:05:44.10006321 +0000 UTC m=+1760.775900104" lastFinishedPulling="2025-12-02 02:05:48.584839919 +0000 UTC m=+1765.260676813" observedRunningTime="2025-12-02 02:05:49.181582404 +0000 UTC m=+1765.857419298" watchObservedRunningTime="2025-12-02 02:05:49.188645416 +0000 UTC m=+1765.864482300" Dec 02 02:05:52 crc kubenswrapper[4884]: I1202 02:05:52.310146 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:05:52 crc kubenswrapper[4884]: I1202 02:05:52.310601 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:05:53 crc kubenswrapper[4884]: I1202 02:05:53.415921 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vj2gg" podUID="f6a2c03c-5d9e-43ec-8b75-63e536b06de5" containerName="registry-server" probeResult="failure" output=< Dec 02 02:05:53 crc kubenswrapper[4884]: timeout: failed to connect service ":50051" within 1s Dec 02 02:05:53 crc kubenswrapper[4884]: > Dec 02 02:05:58 crc kubenswrapper[4884]: I1202 02:05:58.054800 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-b26c-account-create-update-92x4v"] Dec 02 02:05:58 crc kubenswrapper[4884]: I1202 02:05:58.066192 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-05f0-account-create-update-nbjzz"] Dec 02 02:05:58 crc kubenswrapper[4884]: I1202 02:05:58.077672 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-zxwkn"] Dec 02 02:05:58 crc kubenswrapper[4884]: I1202 02:05:58.087204 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-zxwkn"] Dec 02 02:05:58 crc kubenswrapper[4884]: I1202 02:05:58.101716 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-b26c-account-create-update-92x4v"] Dec 02 02:05:58 crc kubenswrapper[4884]: I1202 02:05:58.113332 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-05f0-account-create-update-nbjzz"] Dec 02 02:05:59 crc kubenswrapper[4884]: I1202 02:05:59.053065 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-hnwz5"] Dec 02 02:05:59 crc kubenswrapper[4884]: I1202 02:05:59.065185 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-3040-account-create-update-mx8t8"] Dec 02 02:05:59 crc kubenswrapper[4884]: I1202 02:05:59.076998 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-3040-account-create-update-mx8t8"] Dec 02 02:05:59 crc kubenswrapper[4884]: I1202 02:05:59.089057 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-hnwz5"] Dec 02 02:05:59 crc kubenswrapper[4884]: I1202 02:05:59.096717 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-mrh2r"] Dec 02 02:05:59 crc kubenswrapper[4884]: I1202 02:05:59.104265 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-qfblm"] Dec 02 02:05:59 crc kubenswrapper[4884]: I1202 02:05:59.111428 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-mrh2r"] Dec 02 02:05:59 crc kubenswrapper[4884]: I1202 02:05:59.119270 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-qfblm"] Dec 02 02:05:59 crc kubenswrapper[4884]: I1202 02:05:59.628431 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d14133a-bc13-40b8-9c84-cad4a04bb32e" path="/var/lib/kubelet/pods/3d14133a-bc13-40b8-9c84-cad4a04bb32e/volumes" Dec 02 02:05:59 crc kubenswrapper[4884]: I1202 02:05:59.629562 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e" path="/var/lib/kubelet/pods/83c3e77b-e1e2-4ee7-bd7c-601e4f90ff5e/volumes" Dec 02 02:05:59 crc kubenswrapper[4884]: I1202 02:05:59.631574 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2e81ce0-545f-43bc-b2c2-bea1f931542d" path="/var/lib/kubelet/pods/a2e81ce0-545f-43bc-b2c2-bea1f931542d/volumes" Dec 02 02:05:59 crc kubenswrapper[4884]: I1202 02:05:59.632905 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3baeace-c438-4e75-b503-a6b4ee66e19e" path="/var/lib/kubelet/pods/b3baeace-c438-4e75-b503-a6b4ee66e19e/volumes" Dec 02 02:05:59 crc kubenswrapper[4884]: I1202 02:05:59.634848 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1749692-f3f0-4849-bdf5-73ce5bfec72a" path="/var/lib/kubelet/pods/e1749692-f3f0-4849-bdf5-73ce5bfec72a/volumes" Dec 02 02:05:59 crc kubenswrapper[4884]: I1202 02:05:59.636809 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5722ebb-3612-4608-9768-6d8a95d86efe" path="/var/lib/kubelet/pods/f5722ebb-3612-4608-9768-6d8a95d86efe/volumes" Dec 02 02:05:59 crc kubenswrapper[4884]: I1202 02:05:59.638145 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb68c844-6d7b-4f52-9a57-1ba8a4603449" path="/var/lib/kubelet/pods/fb68c844-6d7b-4f52-9a57-1ba8a4603449/volumes" Dec 02 02:06:01 crc kubenswrapper[4884]: I1202 02:06:01.614633 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:06:01 crc kubenswrapper[4884]: E1202 02:06:01.615381 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:06:02 crc kubenswrapper[4884]: I1202 02:06:02.366547 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:06:02 crc kubenswrapper[4884]: I1202 02:06:02.433316 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:06:02 crc kubenswrapper[4884]: I1202 02:06:02.616962 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vj2gg"] Dec 02 02:06:04 crc kubenswrapper[4884]: I1202 02:06:04.329230 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vj2gg" podUID="f6a2c03c-5d9e-43ec-8b75-63e536b06de5" containerName="registry-server" containerID="cri-o://d941e7beb770c93595314ac8ca0962904699f434a234d88ad6117322fc685905" gracePeriod=2 Dec 02 02:06:04 crc kubenswrapper[4884]: I1202 02:06:04.812144 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:06:04 crc kubenswrapper[4884]: I1202 02:06:04.965077 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-catalog-content\") pod \"f6a2c03c-5d9e-43ec-8b75-63e536b06de5\" (UID: \"f6a2c03c-5d9e-43ec-8b75-63e536b06de5\") " Dec 02 02:06:04 crc kubenswrapper[4884]: I1202 02:06:04.965525 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-utilities\") pod \"f6a2c03c-5d9e-43ec-8b75-63e536b06de5\" (UID: \"f6a2c03c-5d9e-43ec-8b75-63e536b06de5\") " Dec 02 02:06:04 crc kubenswrapper[4884]: I1202 02:06:04.965609 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6j4nw\" (UniqueName: \"kubernetes.io/projected/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-kube-api-access-6j4nw\") pod \"f6a2c03c-5d9e-43ec-8b75-63e536b06de5\" (UID: \"f6a2c03c-5d9e-43ec-8b75-63e536b06de5\") " Dec 02 02:06:04 crc kubenswrapper[4884]: I1202 02:06:04.966933 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-utilities" (OuterVolumeSpecName: "utilities") pod "f6a2c03c-5d9e-43ec-8b75-63e536b06de5" (UID: "f6a2c03c-5d9e-43ec-8b75-63e536b06de5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:06:04 crc kubenswrapper[4884]: I1202 02:06:04.974649 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-kube-api-access-6j4nw" (OuterVolumeSpecName: "kube-api-access-6j4nw") pod "f6a2c03c-5d9e-43ec-8b75-63e536b06de5" (UID: "f6a2c03c-5d9e-43ec-8b75-63e536b06de5"). InnerVolumeSpecName "kube-api-access-6j4nw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.064210 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f6a2c03c-5d9e-43ec-8b75-63e536b06de5" (UID: "f6a2c03c-5d9e-43ec-8b75-63e536b06de5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.068036 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.068066 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.068079 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6j4nw\" (UniqueName: \"kubernetes.io/projected/f6a2c03c-5d9e-43ec-8b75-63e536b06de5-kube-api-access-6j4nw\") on node \"crc\" DevicePath \"\"" Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.346734 4884 generic.go:334] "Generic (PLEG): container finished" podID="f6a2c03c-5d9e-43ec-8b75-63e536b06de5" containerID="d941e7beb770c93595314ac8ca0962904699f434a234d88ad6117322fc685905" exitCode=0 Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.346793 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj2gg" event={"ID":"f6a2c03c-5d9e-43ec-8b75-63e536b06de5","Type":"ContainerDied","Data":"d941e7beb770c93595314ac8ca0962904699f434a234d88ad6117322fc685905"} Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.346825 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj2gg" event={"ID":"f6a2c03c-5d9e-43ec-8b75-63e536b06de5","Type":"ContainerDied","Data":"e190576b67d07a5869af47e97cd2b2c1a1a78f9315f7fa2df6874d9090643d95"} Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.346825 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vj2gg" Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.346844 4884 scope.go:117] "RemoveContainer" containerID="d941e7beb770c93595314ac8ca0962904699f434a234d88ad6117322fc685905" Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.381733 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vj2gg"] Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.385262 4884 scope.go:117] "RemoveContainer" containerID="516ece4494d6076b3c5031c61f71329de801b11c76c8d2b567e2706d2325d06f" Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.393387 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vj2gg"] Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.418953 4884 scope.go:117] "RemoveContainer" containerID="786f216469c9d3e2ef5647cffb0a17791c5402a76953bbfa944cec47dde899cc" Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.468664 4884 scope.go:117] "RemoveContainer" containerID="d941e7beb770c93595314ac8ca0962904699f434a234d88ad6117322fc685905" Dec 02 02:06:05 crc kubenswrapper[4884]: E1202 02:06:05.469162 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d941e7beb770c93595314ac8ca0962904699f434a234d88ad6117322fc685905\": container with ID starting with d941e7beb770c93595314ac8ca0962904699f434a234d88ad6117322fc685905 not found: ID does not exist" containerID="d941e7beb770c93595314ac8ca0962904699f434a234d88ad6117322fc685905" Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.469203 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d941e7beb770c93595314ac8ca0962904699f434a234d88ad6117322fc685905"} err="failed to get container status \"d941e7beb770c93595314ac8ca0962904699f434a234d88ad6117322fc685905\": rpc error: code = NotFound desc = could not find container \"d941e7beb770c93595314ac8ca0962904699f434a234d88ad6117322fc685905\": container with ID starting with d941e7beb770c93595314ac8ca0962904699f434a234d88ad6117322fc685905 not found: ID does not exist" Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.469235 4884 scope.go:117] "RemoveContainer" containerID="516ece4494d6076b3c5031c61f71329de801b11c76c8d2b567e2706d2325d06f" Dec 02 02:06:05 crc kubenswrapper[4884]: E1202 02:06:05.469837 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"516ece4494d6076b3c5031c61f71329de801b11c76c8d2b567e2706d2325d06f\": container with ID starting with 516ece4494d6076b3c5031c61f71329de801b11c76c8d2b567e2706d2325d06f not found: ID does not exist" containerID="516ece4494d6076b3c5031c61f71329de801b11c76c8d2b567e2706d2325d06f" Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.469890 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"516ece4494d6076b3c5031c61f71329de801b11c76c8d2b567e2706d2325d06f"} err="failed to get container status \"516ece4494d6076b3c5031c61f71329de801b11c76c8d2b567e2706d2325d06f\": rpc error: code = NotFound desc = could not find container \"516ece4494d6076b3c5031c61f71329de801b11c76c8d2b567e2706d2325d06f\": container with ID starting with 516ece4494d6076b3c5031c61f71329de801b11c76c8d2b567e2706d2325d06f not found: ID does not exist" Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.469927 4884 scope.go:117] "RemoveContainer" containerID="786f216469c9d3e2ef5647cffb0a17791c5402a76953bbfa944cec47dde899cc" Dec 02 02:06:05 crc kubenswrapper[4884]: E1202 02:06:05.471261 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"786f216469c9d3e2ef5647cffb0a17791c5402a76953bbfa944cec47dde899cc\": container with ID starting with 786f216469c9d3e2ef5647cffb0a17791c5402a76953bbfa944cec47dde899cc not found: ID does not exist" containerID="786f216469c9d3e2ef5647cffb0a17791c5402a76953bbfa944cec47dde899cc" Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.471290 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"786f216469c9d3e2ef5647cffb0a17791c5402a76953bbfa944cec47dde899cc"} err="failed to get container status \"786f216469c9d3e2ef5647cffb0a17791c5402a76953bbfa944cec47dde899cc\": rpc error: code = NotFound desc = could not find container \"786f216469c9d3e2ef5647cffb0a17791c5402a76953bbfa944cec47dde899cc\": container with ID starting with 786f216469c9d3e2ef5647cffb0a17791c5402a76953bbfa944cec47dde899cc not found: ID does not exist" Dec 02 02:06:05 crc kubenswrapper[4884]: I1202 02:06:05.630982 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6a2c03c-5d9e-43ec-8b75-63e536b06de5" path="/var/lib/kubelet/pods/f6a2c03c-5d9e-43ec-8b75-63e536b06de5/volumes" Dec 02 02:06:09 crc kubenswrapper[4884]: I1202 02:06:09.048188 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-6z2v4"] Dec 02 02:06:09 crc kubenswrapper[4884]: I1202 02:06:09.064441 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-6z2v4"] Dec 02 02:06:09 crc kubenswrapper[4884]: I1202 02:06:09.629342 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443" path="/var/lib/kubelet/pods/a8ccc2e4-81c2-434b-a9d4-62fb3e1a4443/volumes" Dec 02 02:06:15 crc kubenswrapper[4884]: I1202 02:06:15.615485 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:06:15 crc kubenswrapper[4884]: E1202 02:06:15.616229 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:06:26 crc kubenswrapper[4884]: I1202 02:06:26.614326 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:06:26 crc kubenswrapper[4884]: E1202 02:06:26.615042 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:06:39 crc kubenswrapper[4884]: I1202 02:06:39.080150 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-jn9fg"] Dec 02 02:06:39 crc kubenswrapper[4884]: I1202 02:06:39.095596 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/watcher-db-sync-pm699"] Dec 02 02:06:39 crc kubenswrapper[4884]: I1202 02:06:39.107225 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/watcher-db-sync-pm699"] Dec 02 02:06:39 crc kubenswrapper[4884]: I1202 02:06:39.117472 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-jn9fg"] Dec 02 02:06:39 crc kubenswrapper[4884]: I1202 02:06:39.615957 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:06:39 crc kubenswrapper[4884]: E1202 02:06:39.616894 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:06:39 crc kubenswrapper[4884]: I1202 02:06:39.630731 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ba1201c-49ba-4c31-bf2a-8069ea8a16c8" path="/var/lib/kubelet/pods/5ba1201c-49ba-4c31-bf2a-8069ea8a16c8/volumes" Dec 02 02:06:39 crc kubenswrapper[4884]: I1202 02:06:39.633630 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d95342e1-6aff-4feb-9309-26ffa69e786b" path="/var/lib/kubelet/pods/d95342e1-6aff-4feb-9309-26ffa69e786b/volumes" Dec 02 02:06:43 crc kubenswrapper[4884]: I1202 02:06:43.803815 4884 scope.go:117] "RemoveContainer" containerID="3a492fb64d69f7c559b62e62078c74a3c29e4de0c8589a224fff85818c397f94" Dec 02 02:06:43 crc kubenswrapper[4884]: I1202 02:06:43.849548 4884 scope.go:117] "RemoveContainer" containerID="92006ba5455afa5d492d493831b1d3ab7855aeeb503b8be1da12487ad1333afe" Dec 02 02:06:43 crc kubenswrapper[4884]: I1202 02:06:43.908234 4884 scope.go:117] "RemoveContainer" containerID="c7fa8f152081d6f4e0e751ec4e3fc7e0aa68e1ac3f86827e68929bc7a0b4adc9" Dec 02 02:06:43 crc kubenswrapper[4884]: I1202 02:06:43.950017 4884 scope.go:117] "RemoveContainer" containerID="5f2b59ff1edac607729bca3d7f71ea5ca3e28819646a4e320756e5f5f9174152" Dec 02 02:06:44 crc kubenswrapper[4884]: I1202 02:06:44.003494 4884 scope.go:117] "RemoveContainer" containerID="20c36ae08413ad701e4df76588767073583a85dac1fbb1d94ea3e0188f80704b" Dec 02 02:06:44 crc kubenswrapper[4884]: I1202 02:06:44.070312 4884 scope.go:117] "RemoveContainer" containerID="f50852b53c911e52af5cc0b9e21d2f0baa40e2bcd63cc95a9dcf9f9f9800b041" Dec 02 02:06:44 crc kubenswrapper[4884]: I1202 02:06:44.101623 4884 scope.go:117] "RemoveContainer" containerID="559d83a50df7830f16bd5fc38db352cb8066dcf9889466f1d0113ad7fd4d74b9" Dec 02 02:06:44 crc kubenswrapper[4884]: I1202 02:06:44.124254 4884 scope.go:117] "RemoveContainer" containerID="5e012e78f633f504286f1a3f8d341b4f23db00e0ffd5bb95de81eb01118a7cfd" Dec 02 02:06:44 crc kubenswrapper[4884]: I1202 02:06:44.148067 4884 scope.go:117] "RemoveContainer" containerID="ccf829018519eb030ae2b299409452e04846a62f68872a2a04271bb793567753" Dec 02 02:06:44 crc kubenswrapper[4884]: I1202 02:06:44.172222 4884 scope.go:117] "RemoveContainer" containerID="a5d68bf6080508333b0425d2b369cbf613c67a868be4a5af9d924e71ccbbd979" Dec 02 02:06:46 crc kubenswrapper[4884]: I1202 02:06:46.044310 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-5vj4w"] Dec 02 02:06:46 crc kubenswrapper[4884]: I1202 02:06:46.051318 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-5vj4w"] Dec 02 02:06:47 crc kubenswrapper[4884]: I1202 02:06:47.627937 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37412014-287b-42f1-9437-ad164aa34f52" path="/var/lib/kubelet/pods/37412014-287b-42f1-9437-ad164aa34f52/volumes" Dec 02 02:06:51 crc kubenswrapper[4884]: I1202 02:06:51.048527 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-wbxwl"] Dec 02 02:06:51 crc kubenswrapper[4884]: I1202 02:06:51.067387 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-wbxwl"] Dec 02 02:06:51 crc kubenswrapper[4884]: I1202 02:06:51.637240 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62db24cb-3633-425d-a4b1-f24b4850ce26" path="/var/lib/kubelet/pods/62db24cb-3633-425d-a4b1-f24b4850ce26/volumes" Dec 02 02:06:54 crc kubenswrapper[4884]: I1202 02:06:54.614805 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:06:54 crc kubenswrapper[4884]: E1202 02:06:54.615431 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:06:58 crc kubenswrapper[4884]: I1202 02:06:58.066131 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-wmhnq"] Dec 02 02:06:58 crc kubenswrapper[4884]: I1202 02:06:58.080127 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-wmhnq"] Dec 02 02:06:59 crc kubenswrapper[4884]: I1202 02:06:59.632949 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe8c221e-e320-4358-b940-436f38a423d3" path="/var/lib/kubelet/pods/fe8c221e-e320-4358-b940-436f38a423d3/volumes" Dec 02 02:07:08 crc kubenswrapper[4884]: I1202 02:07:08.615019 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:07:08 crc kubenswrapper[4884]: E1202 02:07:08.616527 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:07:09 crc kubenswrapper[4884]: I1202 02:07:09.052361 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-c9rsj"] Dec 02 02:07:09 crc kubenswrapper[4884]: I1202 02:07:09.065637 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-c9rsj"] Dec 02 02:07:09 crc kubenswrapper[4884]: I1202 02:07:09.639717 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0d68bd4-6c59-4e41-b32e-f7918d3b4ada" path="/var/lib/kubelet/pods/f0d68bd4-6c59-4e41-b32e-f7918d3b4ada/volumes" Dec 02 02:07:21 crc kubenswrapper[4884]: I1202 02:07:21.614109 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:07:21 crc kubenswrapper[4884]: E1202 02:07:21.615572 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:07:35 crc kubenswrapper[4884]: I1202 02:07:35.614714 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:07:35 crc kubenswrapper[4884]: E1202 02:07:35.615810 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:07:44 crc kubenswrapper[4884]: I1202 02:07:44.490036 4884 scope.go:117] "RemoveContainer" containerID="5a9ade1ccf035898476361d1ac54f2318b834323f15eb54b55c2b391543fcb88" Dec 02 02:07:44 crc kubenswrapper[4884]: I1202 02:07:44.546645 4884 scope.go:117] "RemoveContainer" containerID="1989683bf093eb5ecb39497dbe41e94fa661c162853fe0b1a5047b3edb0cbeb1" Dec 02 02:07:44 crc kubenswrapper[4884]: I1202 02:07:44.596233 4884 scope.go:117] "RemoveContainer" containerID="ea22b91c2dde8cd499c7ebeb771560a39adc369d8ad7edd3dfb5d0effe098cd4" Dec 02 02:07:44 crc kubenswrapper[4884]: I1202 02:07:44.685212 4884 scope.go:117] "RemoveContainer" containerID="decead557aa4909bdaee4b0c705b59f7c4c2cfb3584978e712a87366ff7ab8ac" Dec 02 02:07:50 crc kubenswrapper[4884]: I1202 02:07:50.073239 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-ba12-account-create-update-szvnb"] Dec 02 02:07:50 crc kubenswrapper[4884]: I1202 02:07:50.084952 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-ba12-account-create-update-szvnb"] Dec 02 02:07:50 crc kubenswrapper[4884]: I1202 02:07:50.615082 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:07:50 crc kubenswrapper[4884]: E1202 02:07:50.615474 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.037556 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-hdx8d"] Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.052974 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-1ad5-account-create-update-qw2ms"] Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.065315 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-5nk2w"] Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.073846 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-hdx8d"] Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.082301 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-ee48-account-create-update-bg8ph"] Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.091955 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-1ad5-account-create-update-qw2ms"] Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.098855 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-6d8tj"] Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.106302 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-6d8tj"] Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.113568 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-5nk2w"] Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.121249 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-ee48-account-create-update-bg8ph"] Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.625916 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e7e4b35-7937-45cf-9268-4c510bb9bef7" path="/var/lib/kubelet/pods/2e7e4b35-7937-45cf-9268-4c510bb9bef7/volumes" Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.626971 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3389947b-0d8d-443e-8586-e60740329b9b" path="/var/lib/kubelet/pods/3389947b-0d8d-443e-8586-e60740329b9b/volumes" Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.627824 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ceaef10-69d2-488f-becb-9683a34a59df" path="/var/lib/kubelet/pods/3ceaef10-69d2-488f-becb-9683a34a59df/volumes" Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.628547 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97a8c2ed-ae93-4ad6-944c-beca1829d059" path="/var/lib/kubelet/pods/97a8c2ed-ae93-4ad6-944c-beca1829d059/volumes" Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.630275 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac38729e-6367-4c33-9801-5c79cb0c7fb0" path="/var/lib/kubelet/pods/ac38729e-6367-4c33-9801-5c79cb0c7fb0/volumes" Dec 02 02:07:51 crc kubenswrapper[4884]: I1202 02:07:51.631600 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0" path="/var/lib/kubelet/pods/b3c29ee5-3e34-4de2-b847-4dc4b4c2eed0/volumes" Dec 02 02:08:01 crc kubenswrapper[4884]: I1202 02:08:01.614019 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:08:01 crc kubenswrapper[4884]: E1202 02:08:01.614682 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:08:15 crc kubenswrapper[4884]: I1202 02:08:15.614584 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:08:15 crc kubenswrapper[4884]: E1202 02:08:15.615642 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:08:21 crc kubenswrapper[4884]: I1202 02:08:21.046717 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-f48ww"] Dec 02 02:08:21 crc kubenswrapper[4884]: I1202 02:08:21.056303 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-f48ww"] Dec 02 02:08:21 crc kubenswrapper[4884]: I1202 02:08:21.625283 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26734dfd-6dba-4c5f-99fc-bc27be1b8792" path="/var/lib/kubelet/pods/26734dfd-6dba-4c5f-99fc-bc27be1b8792/volumes" Dec 02 02:08:29 crc kubenswrapper[4884]: I1202 02:08:29.616275 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:08:29 crc kubenswrapper[4884]: E1202 02:08:29.617475 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:08:43 crc kubenswrapper[4884]: I1202 02:08:43.622134 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:08:43 crc kubenswrapper[4884]: E1202 02:08:43.622867 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:08:44 crc kubenswrapper[4884]: I1202 02:08:44.046988 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-kcpzh"] Dec 02 02:08:44 crc kubenswrapper[4884]: I1202 02:08:44.057393 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-x695l"] Dec 02 02:08:44 crc kubenswrapper[4884]: I1202 02:08:44.064992 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-x695l"] Dec 02 02:08:44 crc kubenswrapper[4884]: I1202 02:08:44.091675 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-kcpzh"] Dec 02 02:08:44 crc kubenswrapper[4884]: I1202 02:08:44.858161 4884 scope.go:117] "RemoveContainer" containerID="4dce329c1ffe419e86daff3110d7126e144734dac89421e0b57e19f1763ddaa9" Dec 02 02:08:44 crc kubenswrapper[4884]: I1202 02:08:44.938019 4884 scope.go:117] "RemoveContainer" containerID="8c19950dd25d0e1744f58e97543083aae1f42196c49fb027b5a2876fd7de1a8c" Dec 02 02:08:44 crc kubenswrapper[4884]: I1202 02:08:44.986148 4884 scope.go:117] "RemoveContainer" containerID="225ae74ed43023b6769dfe422299340d8638206e1a4ddcd7cae47f5ba49ada29" Dec 02 02:08:45 crc kubenswrapper[4884]: I1202 02:08:45.041988 4884 scope.go:117] "RemoveContainer" containerID="77e7351dbbd5eabe7d0c4f22fd6464f261cfa31d100e6317bc505e1d5cfa2135" Dec 02 02:08:45 crc kubenswrapper[4884]: I1202 02:08:45.129966 4884 scope.go:117] "RemoveContainer" containerID="621548bbd1d025d4426636ee1e58194771645d67a13795b3461964c8de8c90bb" Dec 02 02:08:45 crc kubenswrapper[4884]: I1202 02:08:45.147250 4884 scope.go:117] "RemoveContainer" containerID="47231ef26f8635d44ac28086955aa3ffd24d73230815cfb6d1cda21d5690f928" Dec 02 02:08:45 crc kubenswrapper[4884]: I1202 02:08:45.209193 4884 scope.go:117] "RemoveContainer" containerID="c9222223659b21deb6248a7babbdceefd8c572f73a4ebf8c62a572b4501bc8f5" Dec 02 02:08:45 crc kubenswrapper[4884]: I1202 02:08:45.629629 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a0a5720-88c1-4adc-9f54-2db975a9e48a" path="/var/lib/kubelet/pods/5a0a5720-88c1-4adc-9f54-2db975a9e48a/volumes" Dec 02 02:08:45 crc kubenswrapper[4884]: I1202 02:08:45.631040 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e74647f6-5b85-47c2-9f26-27465f963095" path="/var/lib/kubelet/pods/e74647f6-5b85-47c2-9f26-27465f963095/volumes" Dec 02 02:08:54 crc kubenswrapper[4884]: I1202 02:08:54.614178 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:08:54 crc kubenswrapper[4884]: E1202 02:08:54.615137 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:09:05 crc kubenswrapper[4884]: I1202 02:09:05.508025 4884 generic.go:334] "Generic (PLEG): container finished" podID="49a3a2e3-b9ff-49bb-b850-5a516270f73b" containerID="2cd97105f2827de7385c5e83ffdcf79bf346a6b42452ecb376cef0b253cff61f" exitCode=0 Dec 02 02:09:05 crc kubenswrapper[4884]: I1202 02:09:05.508145 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" event={"ID":"49a3a2e3-b9ff-49bb-b850-5a516270f73b","Type":"ContainerDied","Data":"2cd97105f2827de7385c5e83ffdcf79bf346a6b42452ecb376cef0b253cff61f"} Dec 02 02:09:06 crc kubenswrapper[4884]: I1202 02:09:06.616302 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:09:06 crc kubenswrapper[4884]: E1202 02:09:06.617003 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.069565 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.201546 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-bootstrap-combined-ca-bundle\") pod \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.201701 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-inventory\") pod \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.201849 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wvq5\" (UniqueName: \"kubernetes.io/projected/49a3a2e3-b9ff-49bb-b850-5a516270f73b-kube-api-access-6wvq5\") pod \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.201938 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-ssh-key\") pod \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\" (UID: \"49a3a2e3-b9ff-49bb-b850-5a516270f73b\") " Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.207617 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "49a3a2e3-b9ff-49bb-b850-5a516270f73b" (UID: "49a3a2e3-b9ff-49bb-b850-5a516270f73b"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.210811 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49a3a2e3-b9ff-49bb-b850-5a516270f73b-kube-api-access-6wvq5" (OuterVolumeSpecName: "kube-api-access-6wvq5") pod "49a3a2e3-b9ff-49bb-b850-5a516270f73b" (UID: "49a3a2e3-b9ff-49bb-b850-5a516270f73b"). InnerVolumeSpecName "kube-api-access-6wvq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.246471 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-inventory" (OuterVolumeSpecName: "inventory") pod "49a3a2e3-b9ff-49bb-b850-5a516270f73b" (UID: "49a3a2e3-b9ff-49bb-b850-5a516270f73b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.257656 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "49a3a2e3-b9ff-49bb-b850-5a516270f73b" (UID: "49a3a2e3-b9ff-49bb-b850-5a516270f73b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.305336 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wvq5\" (UniqueName: \"kubernetes.io/projected/49a3a2e3-b9ff-49bb-b850-5a516270f73b-kube-api-access-6wvq5\") on node \"crc\" DevicePath \"\"" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.305696 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.305716 4884 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.305734 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/49a3a2e3-b9ff-49bb-b850-5a516270f73b-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.531438 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" event={"ID":"49a3a2e3-b9ff-49bb-b850-5a516270f73b","Type":"ContainerDied","Data":"4d0bbf635e647bfbfeb89e001806487438d3f5ecfba1d0839f720d020a478703"} Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.531501 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d0bbf635e647bfbfeb89e001806487438d3f5ecfba1d0839f720d020a478703" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.531515 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.640735 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m"] Dec 02 02:09:07 crc kubenswrapper[4884]: E1202 02:09:07.641098 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49a3a2e3-b9ff-49bb-b850-5a516270f73b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.641112 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="49a3a2e3-b9ff-49bb-b850-5a516270f73b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 02 02:09:07 crc kubenswrapper[4884]: E1202 02:09:07.641124 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6a2c03c-5d9e-43ec-8b75-63e536b06de5" containerName="extract-utilities" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.641130 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6a2c03c-5d9e-43ec-8b75-63e536b06de5" containerName="extract-utilities" Dec 02 02:09:07 crc kubenswrapper[4884]: E1202 02:09:07.641140 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6a2c03c-5d9e-43ec-8b75-63e536b06de5" containerName="extract-content" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.641146 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6a2c03c-5d9e-43ec-8b75-63e536b06de5" containerName="extract-content" Dec 02 02:09:07 crc kubenswrapper[4884]: E1202 02:09:07.641182 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6a2c03c-5d9e-43ec-8b75-63e536b06de5" containerName="registry-server" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.641187 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6a2c03c-5d9e-43ec-8b75-63e536b06de5" containerName="registry-server" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.641357 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="49a3a2e3-b9ff-49bb-b850-5a516270f73b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.641391 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6a2c03c-5d9e-43ec-8b75-63e536b06de5" containerName="registry-server" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.642035 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.666765 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.666955 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.667168 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.667215 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.683621 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m"] Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.715887 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5829ac7-a36e-4c26-a03b-d88518e1361f-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m\" (UID: \"c5829ac7-a36e-4c26-a03b-d88518e1361f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.716266 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxh7p\" (UniqueName: \"kubernetes.io/projected/c5829ac7-a36e-4c26-a03b-d88518e1361f-kube-api-access-cxh7p\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m\" (UID: \"c5829ac7-a36e-4c26-a03b-d88518e1361f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.716349 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5829ac7-a36e-4c26-a03b-d88518e1361f-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m\" (UID: \"c5829ac7-a36e-4c26-a03b-d88518e1361f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.817567 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5829ac7-a36e-4c26-a03b-d88518e1361f-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m\" (UID: \"c5829ac7-a36e-4c26-a03b-d88518e1361f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.817716 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxh7p\" (UniqueName: \"kubernetes.io/projected/c5829ac7-a36e-4c26-a03b-d88518e1361f-kube-api-access-cxh7p\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m\" (UID: \"c5829ac7-a36e-4c26-a03b-d88518e1361f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.817771 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5829ac7-a36e-4c26-a03b-d88518e1361f-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m\" (UID: \"c5829ac7-a36e-4c26-a03b-d88518e1361f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.821925 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5829ac7-a36e-4c26-a03b-d88518e1361f-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m\" (UID: \"c5829ac7-a36e-4c26-a03b-d88518e1361f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.822849 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5829ac7-a36e-4c26-a03b-d88518e1361f-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m\" (UID: \"c5829ac7-a36e-4c26-a03b-d88518e1361f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.836719 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxh7p\" (UniqueName: \"kubernetes.io/projected/c5829ac7-a36e-4c26-a03b-d88518e1361f-kube-api-access-cxh7p\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m\" (UID: \"c5829ac7-a36e-4c26-a03b-d88518e1361f\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" Dec 02 02:09:07 crc kubenswrapper[4884]: I1202 02:09:07.985657 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" Dec 02 02:09:08 crc kubenswrapper[4884]: I1202 02:09:08.567678 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m"] Dec 02 02:09:08 crc kubenswrapper[4884]: W1202 02:09:08.570971 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5829ac7_a36e_4c26_a03b_d88518e1361f.slice/crio-0ba838e80039b4b2bc9af9ec5142f05956d5af8ac9409e846edd8b1d419e5286 WatchSource:0}: Error finding container 0ba838e80039b4b2bc9af9ec5142f05956d5af8ac9409e846edd8b1d419e5286: Status 404 returned error can't find the container with id 0ba838e80039b4b2bc9af9ec5142f05956d5af8ac9409e846edd8b1d419e5286 Dec 02 02:09:09 crc kubenswrapper[4884]: I1202 02:09:09.551839 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" event={"ID":"c5829ac7-a36e-4c26-a03b-d88518e1361f","Type":"ContainerStarted","Data":"b75111785fde9a9ef89087de1c16938f9252f0776dae7b85f0ec9bb4e6385950"} Dec 02 02:09:09 crc kubenswrapper[4884]: I1202 02:09:09.552203 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" event={"ID":"c5829ac7-a36e-4c26-a03b-d88518e1361f","Type":"ContainerStarted","Data":"0ba838e80039b4b2bc9af9ec5142f05956d5af8ac9409e846edd8b1d419e5286"} Dec 02 02:09:09 crc kubenswrapper[4884]: I1202 02:09:09.580290 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" podStartSLOduration=1.9802173760000001 podStartE2EDuration="2.580271236s" podCreationTimestamp="2025-12-02 02:09:07 +0000 UTC" firstStartedPulling="2025-12-02 02:09:08.574785459 +0000 UTC m=+1965.250622383" lastFinishedPulling="2025-12-02 02:09:09.174839319 +0000 UTC m=+1965.850676243" observedRunningTime="2025-12-02 02:09:09.569731367 +0000 UTC m=+1966.245568271" watchObservedRunningTime="2025-12-02 02:09:09.580271236 +0000 UTC m=+1966.256108130" Dec 02 02:09:18 crc kubenswrapper[4884]: I1202 02:09:18.614083 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:09:19 crc kubenswrapper[4884]: I1202 02:09:19.685386 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"c0641b565628601f6c7f6d17e6aebb25f493b176f7f04ae6e89252cccd560d8e"} Dec 02 02:09:29 crc kubenswrapper[4884]: I1202 02:09:29.060364 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-djwlt"] Dec 02 02:09:29 crc kubenswrapper[4884]: I1202 02:09:29.071997 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-djwlt"] Dec 02 02:09:29 crc kubenswrapper[4884]: I1202 02:09:29.637266 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1acb68f2-bed9-462f-82eb-244f84069402" path="/var/lib/kubelet/pods/1acb68f2-bed9-462f-82eb-244f84069402/volumes" Dec 02 02:09:45 crc kubenswrapper[4884]: I1202 02:09:45.361003 4884 scope.go:117] "RemoveContainer" containerID="a4b47a3ed380d29bda07b1cc310be577ed09eece497e9ca981028164b29fc5a4" Dec 02 02:09:45 crc kubenswrapper[4884]: I1202 02:09:45.418898 4884 scope.go:117] "RemoveContainer" containerID="e54c5930d9ab03a015536b995f4926f387f59c08dd7848b2af7b37b877f67176" Dec 02 02:09:45 crc kubenswrapper[4884]: I1202 02:09:45.483215 4884 scope.go:117] "RemoveContainer" containerID="ab1060d0fe9f461502ed0a14aecefbdef801e68eae01d8e5e79d6aa38891af6d" Dec 02 02:11:34 crc kubenswrapper[4884]: I1202 02:11:34.240289 4884 generic.go:334] "Generic (PLEG): container finished" podID="c5829ac7-a36e-4c26-a03b-d88518e1361f" containerID="b75111785fde9a9ef89087de1c16938f9252f0776dae7b85f0ec9bb4e6385950" exitCode=0 Dec 02 02:11:34 crc kubenswrapper[4884]: I1202 02:11:34.240401 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" event={"ID":"c5829ac7-a36e-4c26-a03b-d88518e1361f","Type":"ContainerDied","Data":"b75111785fde9a9ef89087de1c16938f9252f0776dae7b85f0ec9bb4e6385950"} Dec 02 02:11:35 crc kubenswrapper[4884]: I1202 02:11:35.720206 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" Dec 02 02:11:35 crc kubenswrapper[4884]: I1202 02:11:35.813719 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxh7p\" (UniqueName: \"kubernetes.io/projected/c5829ac7-a36e-4c26-a03b-d88518e1361f-kube-api-access-cxh7p\") pod \"c5829ac7-a36e-4c26-a03b-d88518e1361f\" (UID: \"c5829ac7-a36e-4c26-a03b-d88518e1361f\") " Dec 02 02:11:35 crc kubenswrapper[4884]: I1202 02:11:35.813897 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5829ac7-a36e-4c26-a03b-d88518e1361f-ssh-key\") pod \"c5829ac7-a36e-4c26-a03b-d88518e1361f\" (UID: \"c5829ac7-a36e-4c26-a03b-d88518e1361f\") " Dec 02 02:11:35 crc kubenswrapper[4884]: I1202 02:11:35.813954 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5829ac7-a36e-4c26-a03b-d88518e1361f-inventory\") pod \"c5829ac7-a36e-4c26-a03b-d88518e1361f\" (UID: \"c5829ac7-a36e-4c26-a03b-d88518e1361f\") " Dec 02 02:11:35 crc kubenswrapper[4884]: I1202 02:11:35.825579 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5829ac7-a36e-4c26-a03b-d88518e1361f-kube-api-access-cxh7p" (OuterVolumeSpecName: "kube-api-access-cxh7p") pod "c5829ac7-a36e-4c26-a03b-d88518e1361f" (UID: "c5829ac7-a36e-4c26-a03b-d88518e1361f"). InnerVolumeSpecName "kube-api-access-cxh7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:11:35 crc kubenswrapper[4884]: I1202 02:11:35.859621 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5829ac7-a36e-4c26-a03b-d88518e1361f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c5829ac7-a36e-4c26-a03b-d88518e1361f" (UID: "c5829ac7-a36e-4c26-a03b-d88518e1361f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:11:35 crc kubenswrapper[4884]: I1202 02:11:35.862345 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5829ac7-a36e-4c26-a03b-d88518e1361f-inventory" (OuterVolumeSpecName: "inventory") pod "c5829ac7-a36e-4c26-a03b-d88518e1361f" (UID: "c5829ac7-a36e-4c26-a03b-d88518e1361f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:11:35 crc kubenswrapper[4884]: I1202 02:11:35.917581 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxh7p\" (UniqueName: \"kubernetes.io/projected/c5829ac7-a36e-4c26-a03b-d88518e1361f-kube-api-access-cxh7p\") on node \"crc\" DevicePath \"\"" Dec 02 02:11:35 crc kubenswrapper[4884]: I1202 02:11:35.919203 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c5829ac7-a36e-4c26-a03b-d88518e1361f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:11:35 crc kubenswrapper[4884]: I1202 02:11:35.919979 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c5829ac7-a36e-4c26-a03b-d88518e1361f-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.270603 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" event={"ID":"c5829ac7-a36e-4c26-a03b-d88518e1361f","Type":"ContainerDied","Data":"0ba838e80039b4b2bc9af9ec5142f05956d5af8ac9409e846edd8b1d419e5286"} Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.270676 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ba838e80039b4b2bc9af9ec5142f05956d5af8ac9409e846edd8b1d419e5286" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.270680 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.423635 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2"] Dec 02 02:11:36 crc kubenswrapper[4884]: E1202 02:11:36.424262 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5829ac7-a36e-4c26-a03b-d88518e1361f" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.424295 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5829ac7-a36e-4c26-a03b-d88518e1361f" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.424659 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5829ac7-a36e-4c26-a03b-d88518e1361f" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.425913 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.428548 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.428962 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.430972 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.431399 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.439041 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2"] Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.532684 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pss74\" (UniqueName: \"kubernetes.io/projected/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-kube-api-access-pss74\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4gns2\" (UID: \"c129a2ce-c6ab-47b5-af8f-55dea1e97d90\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.533212 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4gns2\" (UID: \"c129a2ce-c6ab-47b5-af8f-55dea1e97d90\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.533824 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4gns2\" (UID: \"c129a2ce-c6ab-47b5-af8f-55dea1e97d90\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.635358 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pss74\" (UniqueName: \"kubernetes.io/projected/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-kube-api-access-pss74\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4gns2\" (UID: \"c129a2ce-c6ab-47b5-af8f-55dea1e97d90\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.635804 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4gns2\" (UID: \"c129a2ce-c6ab-47b5-af8f-55dea1e97d90\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.635929 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4gns2\" (UID: \"c129a2ce-c6ab-47b5-af8f-55dea1e97d90\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.643188 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4gns2\" (UID: \"c129a2ce-c6ab-47b5-af8f-55dea1e97d90\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.646444 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4gns2\" (UID: \"c129a2ce-c6ab-47b5-af8f-55dea1e97d90\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.669073 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pss74\" (UniqueName: \"kubernetes.io/projected/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-kube-api-access-pss74\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4gns2\" (UID: \"c129a2ce-c6ab-47b5-af8f-55dea1e97d90\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" Dec 02 02:11:36 crc kubenswrapper[4884]: I1202 02:11:36.767117 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" Dec 02 02:11:37 crc kubenswrapper[4884]: I1202 02:11:37.417621 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2"] Dec 02 02:11:37 crc kubenswrapper[4884]: W1202 02:11:37.429288 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc129a2ce_c6ab_47b5_af8f_55dea1e97d90.slice/crio-23bebc80e2d3d7eeecb9e83470566463cd73430e99310346f329c69d7d8777c0 WatchSource:0}: Error finding container 23bebc80e2d3d7eeecb9e83470566463cd73430e99310346f329c69d7d8777c0: Status 404 returned error can't find the container with id 23bebc80e2d3d7eeecb9e83470566463cd73430e99310346f329c69d7d8777c0 Dec 02 02:11:37 crc kubenswrapper[4884]: I1202 02:11:37.432930 4884 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 02:11:38 crc kubenswrapper[4884]: I1202 02:11:38.303283 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" event={"ID":"c129a2ce-c6ab-47b5-af8f-55dea1e97d90","Type":"ContainerStarted","Data":"877b9757443a6983258b6ec383d20845859e8874ee48bb48dd4794b2b9ec21e8"} Dec 02 02:11:38 crc kubenswrapper[4884]: I1202 02:11:38.303630 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" event={"ID":"c129a2ce-c6ab-47b5-af8f-55dea1e97d90","Type":"ContainerStarted","Data":"23bebc80e2d3d7eeecb9e83470566463cd73430e99310346f329c69d7d8777c0"} Dec 02 02:11:38 crc kubenswrapper[4884]: I1202 02:11:38.327876 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" podStartSLOduration=1.724750964 podStartE2EDuration="2.32785016s" podCreationTimestamp="2025-12-02 02:11:36 +0000 UTC" firstStartedPulling="2025-12-02 02:11:37.432481819 +0000 UTC m=+2114.108318733" lastFinishedPulling="2025-12-02 02:11:38.035581035 +0000 UTC m=+2114.711417929" observedRunningTime="2025-12-02 02:11:38.321404931 +0000 UTC m=+2114.997241825" watchObservedRunningTime="2025-12-02 02:11:38.32785016 +0000 UTC m=+2115.003687094" Dec 02 02:11:46 crc kubenswrapper[4884]: I1202 02:11:46.972278 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:11:46 crc kubenswrapper[4884]: I1202 02:11:46.973077 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:12:12 crc kubenswrapper[4884]: I1202 02:12:12.353488 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rl7gz"] Dec 02 02:12:12 crc kubenswrapper[4884]: I1202 02:12:12.361627 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:12 crc kubenswrapper[4884]: I1202 02:12:12.391903 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-utilities\") pod \"certified-operators-rl7gz\" (UID: \"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db\") " pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:12 crc kubenswrapper[4884]: I1202 02:12:12.391985 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-catalog-content\") pod \"certified-operators-rl7gz\" (UID: \"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db\") " pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:12 crc kubenswrapper[4884]: I1202 02:12:12.392034 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz2gp\" (UniqueName: \"kubernetes.io/projected/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-kube-api-access-pz2gp\") pod \"certified-operators-rl7gz\" (UID: \"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db\") " pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:12 crc kubenswrapper[4884]: I1202 02:12:12.400826 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rl7gz"] Dec 02 02:12:12 crc kubenswrapper[4884]: I1202 02:12:12.495331 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-utilities\") pod \"certified-operators-rl7gz\" (UID: \"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db\") " pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:12 crc kubenswrapper[4884]: I1202 02:12:12.495562 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-catalog-content\") pod \"certified-operators-rl7gz\" (UID: \"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db\") " pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:12 crc kubenswrapper[4884]: I1202 02:12:12.495629 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz2gp\" (UniqueName: \"kubernetes.io/projected/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-kube-api-access-pz2gp\") pod \"certified-operators-rl7gz\" (UID: \"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db\") " pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:12 crc kubenswrapper[4884]: I1202 02:12:12.496170 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-utilities\") pod \"certified-operators-rl7gz\" (UID: \"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db\") " pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:12 crc kubenswrapper[4884]: I1202 02:12:12.496531 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-catalog-content\") pod \"certified-operators-rl7gz\" (UID: \"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db\") " pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:12 crc kubenswrapper[4884]: I1202 02:12:12.537303 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz2gp\" (UniqueName: \"kubernetes.io/projected/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-kube-api-access-pz2gp\") pod \"certified-operators-rl7gz\" (UID: \"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db\") " pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:12 crc kubenswrapper[4884]: I1202 02:12:12.687158 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:12 crc kubenswrapper[4884]: I1202 02:12:12.949357 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rl7gz"] Dec 02 02:12:13 crc kubenswrapper[4884]: I1202 02:12:13.707116 4884 generic.go:334] "Generic (PLEG): container finished" podID="cf78e4c5-441b-4ebe-86f7-0e0edd82a5db" containerID="f977e797e27b8de96edef76882167e2c5b452e425a530e1adfc475fe03491db7" exitCode=0 Dec 02 02:12:13 crc kubenswrapper[4884]: I1202 02:12:13.707576 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rl7gz" event={"ID":"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db","Type":"ContainerDied","Data":"f977e797e27b8de96edef76882167e2c5b452e425a530e1adfc475fe03491db7"} Dec 02 02:12:13 crc kubenswrapper[4884]: I1202 02:12:13.707667 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rl7gz" event={"ID":"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db","Type":"ContainerStarted","Data":"e7bedcba61b7e8337ab924e88d2a2dde2d1015c45d91c3e2b9470e2f65efa5fc"} Dec 02 02:12:14 crc kubenswrapper[4884]: I1202 02:12:14.726620 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rl7gz" event={"ID":"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db","Type":"ContainerStarted","Data":"0d632087dbd3b69d7ac12b70ca34894c3cea3d4c7e3738031163c8ba7e96e6c4"} Dec 02 02:12:15 crc kubenswrapper[4884]: I1202 02:12:15.742163 4884 generic.go:334] "Generic (PLEG): container finished" podID="cf78e4c5-441b-4ebe-86f7-0e0edd82a5db" containerID="0d632087dbd3b69d7ac12b70ca34894c3cea3d4c7e3738031163c8ba7e96e6c4" exitCode=0 Dec 02 02:12:15 crc kubenswrapper[4884]: I1202 02:12:15.742308 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rl7gz" event={"ID":"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db","Type":"ContainerDied","Data":"0d632087dbd3b69d7ac12b70ca34894c3cea3d4c7e3738031163c8ba7e96e6c4"} Dec 02 02:12:16 crc kubenswrapper[4884]: I1202 02:12:16.754619 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rl7gz" event={"ID":"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db","Type":"ContainerStarted","Data":"74bd9d294275a6854bc549fbe7135a8f732138d357d734424c41700a7f9d780a"} Dec 02 02:12:16 crc kubenswrapper[4884]: I1202 02:12:16.790108 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rl7gz" podStartSLOduration=2.07990327 podStartE2EDuration="4.790086704s" podCreationTimestamp="2025-12-02 02:12:12 +0000 UTC" firstStartedPulling="2025-12-02 02:12:13.709966786 +0000 UTC m=+2150.385803680" lastFinishedPulling="2025-12-02 02:12:16.42015019 +0000 UTC m=+2153.095987114" observedRunningTime="2025-12-02 02:12:16.779293348 +0000 UTC m=+2153.455130262" watchObservedRunningTime="2025-12-02 02:12:16.790086704 +0000 UTC m=+2153.465923598" Dec 02 02:12:16 crc kubenswrapper[4884]: I1202 02:12:16.971612 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:12:16 crc kubenswrapper[4884]: I1202 02:12:16.971895 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:12:22 crc kubenswrapper[4884]: I1202 02:12:22.688306 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:22 crc kubenswrapper[4884]: I1202 02:12:22.688862 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:22 crc kubenswrapper[4884]: I1202 02:12:22.780799 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:22 crc kubenswrapper[4884]: I1202 02:12:22.906648 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:23 crc kubenswrapper[4884]: I1202 02:12:23.032165 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rl7gz"] Dec 02 02:12:24 crc kubenswrapper[4884]: I1202 02:12:24.865531 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rl7gz" podUID="cf78e4c5-441b-4ebe-86f7-0e0edd82a5db" containerName="registry-server" containerID="cri-o://74bd9d294275a6854bc549fbe7135a8f732138d357d734424c41700a7f9d780a" gracePeriod=2 Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.447840 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.488343 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pz2gp\" (UniqueName: \"kubernetes.io/projected/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-kube-api-access-pz2gp\") pod \"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db\" (UID: \"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db\") " Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.488583 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-catalog-content\") pod \"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db\" (UID: \"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db\") " Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.488682 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-utilities\") pod \"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db\" (UID: \"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db\") " Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.490304 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-utilities" (OuterVolumeSpecName: "utilities") pod "cf78e4c5-441b-4ebe-86f7-0e0edd82a5db" (UID: "cf78e4c5-441b-4ebe-86f7-0e0edd82a5db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.498807 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-kube-api-access-pz2gp" (OuterVolumeSpecName: "kube-api-access-pz2gp") pod "cf78e4c5-441b-4ebe-86f7-0e0edd82a5db" (UID: "cf78e4c5-441b-4ebe-86f7-0e0edd82a5db"). InnerVolumeSpecName "kube-api-access-pz2gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.590734 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.591164 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz2gp\" (UniqueName: \"kubernetes.io/projected/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-kube-api-access-pz2gp\") on node \"crc\" DevicePath \"\"" Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.796791 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf78e4c5-441b-4ebe-86f7-0e0edd82a5db" (UID: "cf78e4c5-441b-4ebe-86f7-0e0edd82a5db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.879988 4884 generic.go:334] "Generic (PLEG): container finished" podID="cf78e4c5-441b-4ebe-86f7-0e0edd82a5db" containerID="74bd9d294275a6854bc549fbe7135a8f732138d357d734424c41700a7f9d780a" exitCode=0 Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.880056 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rl7gz" event={"ID":"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db","Type":"ContainerDied","Data":"74bd9d294275a6854bc549fbe7135a8f732138d357d734424c41700a7f9d780a"} Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.880098 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rl7gz" event={"ID":"cf78e4c5-441b-4ebe-86f7-0e0edd82a5db","Type":"ContainerDied","Data":"e7bedcba61b7e8337ab924e88d2a2dde2d1015c45d91c3e2b9470e2f65efa5fc"} Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.880128 4884 scope.go:117] "RemoveContainer" containerID="74bd9d294275a6854bc549fbe7135a8f732138d357d734424c41700a7f9d780a" Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.880171 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rl7gz" Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.898624 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.912990 4884 scope.go:117] "RemoveContainer" containerID="0d632087dbd3b69d7ac12b70ca34894c3cea3d4c7e3738031163c8ba7e96e6c4" Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.935037 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rl7gz"] Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.942661 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rl7gz"] Dec 02 02:12:25 crc kubenswrapper[4884]: I1202 02:12:25.959330 4884 scope.go:117] "RemoveContainer" containerID="f977e797e27b8de96edef76882167e2c5b452e425a530e1adfc475fe03491db7" Dec 02 02:12:26 crc kubenswrapper[4884]: I1202 02:12:26.007108 4884 scope.go:117] "RemoveContainer" containerID="74bd9d294275a6854bc549fbe7135a8f732138d357d734424c41700a7f9d780a" Dec 02 02:12:26 crc kubenswrapper[4884]: E1202 02:12:26.008374 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74bd9d294275a6854bc549fbe7135a8f732138d357d734424c41700a7f9d780a\": container with ID starting with 74bd9d294275a6854bc549fbe7135a8f732138d357d734424c41700a7f9d780a not found: ID does not exist" containerID="74bd9d294275a6854bc549fbe7135a8f732138d357d734424c41700a7f9d780a" Dec 02 02:12:26 crc kubenswrapper[4884]: I1202 02:12:26.008419 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74bd9d294275a6854bc549fbe7135a8f732138d357d734424c41700a7f9d780a"} err="failed to get container status \"74bd9d294275a6854bc549fbe7135a8f732138d357d734424c41700a7f9d780a\": rpc error: code = NotFound desc = could not find container \"74bd9d294275a6854bc549fbe7135a8f732138d357d734424c41700a7f9d780a\": container with ID starting with 74bd9d294275a6854bc549fbe7135a8f732138d357d734424c41700a7f9d780a not found: ID does not exist" Dec 02 02:12:26 crc kubenswrapper[4884]: I1202 02:12:26.008455 4884 scope.go:117] "RemoveContainer" containerID="0d632087dbd3b69d7ac12b70ca34894c3cea3d4c7e3738031163c8ba7e96e6c4" Dec 02 02:12:26 crc kubenswrapper[4884]: E1202 02:12:26.009148 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d632087dbd3b69d7ac12b70ca34894c3cea3d4c7e3738031163c8ba7e96e6c4\": container with ID starting with 0d632087dbd3b69d7ac12b70ca34894c3cea3d4c7e3738031163c8ba7e96e6c4 not found: ID does not exist" containerID="0d632087dbd3b69d7ac12b70ca34894c3cea3d4c7e3738031163c8ba7e96e6c4" Dec 02 02:12:26 crc kubenswrapper[4884]: I1202 02:12:26.009189 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d632087dbd3b69d7ac12b70ca34894c3cea3d4c7e3738031163c8ba7e96e6c4"} err="failed to get container status \"0d632087dbd3b69d7ac12b70ca34894c3cea3d4c7e3738031163c8ba7e96e6c4\": rpc error: code = NotFound desc = could not find container \"0d632087dbd3b69d7ac12b70ca34894c3cea3d4c7e3738031163c8ba7e96e6c4\": container with ID starting with 0d632087dbd3b69d7ac12b70ca34894c3cea3d4c7e3738031163c8ba7e96e6c4 not found: ID does not exist" Dec 02 02:12:26 crc kubenswrapper[4884]: I1202 02:12:26.009207 4884 scope.go:117] "RemoveContainer" containerID="f977e797e27b8de96edef76882167e2c5b452e425a530e1adfc475fe03491db7" Dec 02 02:12:26 crc kubenswrapper[4884]: E1202 02:12:26.009554 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f977e797e27b8de96edef76882167e2c5b452e425a530e1adfc475fe03491db7\": container with ID starting with f977e797e27b8de96edef76882167e2c5b452e425a530e1adfc475fe03491db7 not found: ID does not exist" containerID="f977e797e27b8de96edef76882167e2c5b452e425a530e1adfc475fe03491db7" Dec 02 02:12:26 crc kubenswrapper[4884]: I1202 02:12:26.009603 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f977e797e27b8de96edef76882167e2c5b452e425a530e1adfc475fe03491db7"} err="failed to get container status \"f977e797e27b8de96edef76882167e2c5b452e425a530e1adfc475fe03491db7\": rpc error: code = NotFound desc = could not find container \"f977e797e27b8de96edef76882167e2c5b452e425a530e1adfc475fe03491db7\": container with ID starting with f977e797e27b8de96edef76882167e2c5b452e425a530e1adfc475fe03491db7 not found: ID does not exist" Dec 02 02:12:27 crc kubenswrapper[4884]: I1202 02:12:27.630003 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf78e4c5-441b-4ebe-86f7-0e0edd82a5db" path="/var/lib/kubelet/pods/cf78e4c5-441b-4ebe-86f7-0e0edd82a5db/volumes" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.437503 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g45qg"] Dec 02 02:12:34 crc kubenswrapper[4884]: E1202 02:12:34.438343 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf78e4c5-441b-4ebe-86f7-0e0edd82a5db" containerName="registry-server" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.438357 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf78e4c5-441b-4ebe-86f7-0e0edd82a5db" containerName="registry-server" Dec 02 02:12:34 crc kubenswrapper[4884]: E1202 02:12:34.438377 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf78e4c5-441b-4ebe-86f7-0e0edd82a5db" containerName="extract-content" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.438384 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf78e4c5-441b-4ebe-86f7-0e0edd82a5db" containerName="extract-content" Dec 02 02:12:34 crc kubenswrapper[4884]: E1202 02:12:34.438397 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf78e4c5-441b-4ebe-86f7-0e0edd82a5db" containerName="extract-utilities" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.438404 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf78e4c5-441b-4ebe-86f7-0e0edd82a5db" containerName="extract-utilities" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.438605 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf78e4c5-441b-4ebe-86f7-0e0edd82a5db" containerName="registry-server" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.439956 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.474444 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g45qg"] Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.606476 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-utilities\") pod \"community-operators-g45qg\" (UID: \"d8ecb249-9a8d-442a-86ae-d4d28bd3af21\") " pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.606583 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d77h6\" (UniqueName: \"kubernetes.io/projected/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-kube-api-access-d77h6\") pod \"community-operators-g45qg\" (UID: \"d8ecb249-9a8d-442a-86ae-d4d28bd3af21\") " pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.606656 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-catalog-content\") pod \"community-operators-g45qg\" (UID: \"d8ecb249-9a8d-442a-86ae-d4d28bd3af21\") " pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.640409 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6tgs2"] Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.643240 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.690794 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tgs2"] Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.709989 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-utilities\") pod \"redhat-marketplace-6tgs2\" (UID: \"8bf42fe2-0faf-480f-baa0-5a987e80c4f0\") " pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.710044 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qlgb\" (UniqueName: \"kubernetes.io/projected/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-kube-api-access-4qlgb\") pod \"redhat-marketplace-6tgs2\" (UID: \"8bf42fe2-0faf-480f-baa0-5a987e80c4f0\") " pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.710098 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-utilities\") pod \"community-operators-g45qg\" (UID: \"d8ecb249-9a8d-442a-86ae-d4d28bd3af21\") " pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.710140 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-catalog-content\") pod \"redhat-marketplace-6tgs2\" (UID: \"8bf42fe2-0faf-480f-baa0-5a987e80c4f0\") " pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.710176 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d77h6\" (UniqueName: \"kubernetes.io/projected/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-kube-api-access-d77h6\") pod \"community-operators-g45qg\" (UID: \"d8ecb249-9a8d-442a-86ae-d4d28bd3af21\") " pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.710568 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-catalog-content\") pod \"community-operators-g45qg\" (UID: \"d8ecb249-9a8d-442a-86ae-d4d28bd3af21\") " pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.710603 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-utilities\") pod \"community-operators-g45qg\" (UID: \"d8ecb249-9a8d-442a-86ae-d4d28bd3af21\") " pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.711009 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-catalog-content\") pod \"community-operators-g45qg\" (UID: \"d8ecb249-9a8d-442a-86ae-d4d28bd3af21\") " pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.737213 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d77h6\" (UniqueName: \"kubernetes.io/projected/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-kube-api-access-d77h6\") pod \"community-operators-g45qg\" (UID: \"d8ecb249-9a8d-442a-86ae-d4d28bd3af21\") " pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.779312 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.817094 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-catalog-content\") pod \"redhat-marketplace-6tgs2\" (UID: \"8bf42fe2-0faf-480f-baa0-5a987e80c4f0\") " pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.817308 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-utilities\") pod \"redhat-marketplace-6tgs2\" (UID: \"8bf42fe2-0faf-480f-baa0-5a987e80c4f0\") " pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.817337 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qlgb\" (UniqueName: \"kubernetes.io/projected/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-kube-api-access-4qlgb\") pod \"redhat-marketplace-6tgs2\" (UID: \"8bf42fe2-0faf-480f-baa0-5a987e80c4f0\") " pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.817681 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-catalog-content\") pod \"redhat-marketplace-6tgs2\" (UID: \"8bf42fe2-0faf-480f-baa0-5a987e80c4f0\") " pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.817960 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-utilities\") pod \"redhat-marketplace-6tgs2\" (UID: \"8bf42fe2-0faf-480f-baa0-5a987e80c4f0\") " pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.839442 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qlgb\" (UniqueName: \"kubernetes.io/projected/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-kube-api-access-4qlgb\") pod \"redhat-marketplace-6tgs2\" (UID: \"8bf42fe2-0faf-480f-baa0-5a987e80c4f0\") " pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:34 crc kubenswrapper[4884]: I1202 02:12:34.971954 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:35 crc kubenswrapper[4884]: I1202 02:12:35.130262 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g45qg"] Dec 02 02:12:35 crc kubenswrapper[4884]: W1202 02:12:35.600249 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bf42fe2_0faf_480f_baa0_5a987e80c4f0.slice/crio-5ac074ee2d19eaa9e8f19f3dedfc5e888eb6378abda26cd434084a40730b1c3e WatchSource:0}: Error finding container 5ac074ee2d19eaa9e8f19f3dedfc5e888eb6378abda26cd434084a40730b1c3e: Status 404 returned error can't find the container with id 5ac074ee2d19eaa9e8f19f3dedfc5e888eb6378abda26cd434084a40730b1c3e Dec 02 02:12:35 crc kubenswrapper[4884]: I1202 02:12:35.606587 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tgs2"] Dec 02 02:12:36 crc kubenswrapper[4884]: I1202 02:12:36.028053 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8ecb249-9a8d-442a-86ae-d4d28bd3af21" containerID="e883433eb2bfaf2913ed00212542355e93def7a0cbde5d527d0c3ed54f3f2cd5" exitCode=0 Dec 02 02:12:36 crc kubenswrapper[4884]: I1202 02:12:36.028113 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g45qg" event={"ID":"d8ecb249-9a8d-442a-86ae-d4d28bd3af21","Type":"ContainerDied","Data":"e883433eb2bfaf2913ed00212542355e93def7a0cbde5d527d0c3ed54f3f2cd5"} Dec 02 02:12:36 crc kubenswrapper[4884]: I1202 02:12:36.028352 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g45qg" event={"ID":"d8ecb249-9a8d-442a-86ae-d4d28bd3af21","Type":"ContainerStarted","Data":"8047505ac4deb0ba78ea434be0ff9cbfbd57b6b666ad0d77035d897f8e538430"} Dec 02 02:12:36 crc kubenswrapper[4884]: I1202 02:12:36.032493 4884 generic.go:334] "Generic (PLEG): container finished" podID="8bf42fe2-0faf-480f-baa0-5a987e80c4f0" containerID="e582571a7febc3149bc9ebeb07e245ea3e3e56ab24ac2064c234a6e07da22144" exitCode=0 Dec 02 02:12:36 crc kubenswrapper[4884]: I1202 02:12:36.032547 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tgs2" event={"ID":"8bf42fe2-0faf-480f-baa0-5a987e80c4f0","Type":"ContainerDied","Data":"e582571a7febc3149bc9ebeb07e245ea3e3e56ab24ac2064c234a6e07da22144"} Dec 02 02:12:36 crc kubenswrapper[4884]: I1202 02:12:36.032578 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tgs2" event={"ID":"8bf42fe2-0faf-480f-baa0-5a987e80c4f0","Type":"ContainerStarted","Data":"5ac074ee2d19eaa9e8f19f3dedfc5e888eb6378abda26cd434084a40730b1c3e"} Dec 02 02:12:37 crc kubenswrapper[4884]: I1202 02:12:37.047198 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tgs2" event={"ID":"8bf42fe2-0faf-480f-baa0-5a987e80c4f0","Type":"ContainerStarted","Data":"384f40eabd18712e702e55e7d8fe0183e8b055078453e73ba8ab941aed36a1e4"} Dec 02 02:12:38 crc kubenswrapper[4884]: I1202 02:12:38.064335 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8ecb249-9a8d-442a-86ae-d4d28bd3af21" containerID="82ba5c1a84c691497c8ed8d1be3caed480647326b4ee53dbd1de911f5787de26" exitCode=0 Dec 02 02:12:38 crc kubenswrapper[4884]: I1202 02:12:38.064462 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g45qg" event={"ID":"d8ecb249-9a8d-442a-86ae-d4d28bd3af21","Type":"ContainerDied","Data":"82ba5c1a84c691497c8ed8d1be3caed480647326b4ee53dbd1de911f5787de26"} Dec 02 02:12:38 crc kubenswrapper[4884]: I1202 02:12:38.069472 4884 generic.go:334] "Generic (PLEG): container finished" podID="8bf42fe2-0faf-480f-baa0-5a987e80c4f0" containerID="384f40eabd18712e702e55e7d8fe0183e8b055078453e73ba8ab941aed36a1e4" exitCode=0 Dec 02 02:12:38 crc kubenswrapper[4884]: I1202 02:12:38.069511 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tgs2" event={"ID":"8bf42fe2-0faf-480f-baa0-5a987e80c4f0","Type":"ContainerDied","Data":"384f40eabd18712e702e55e7d8fe0183e8b055078453e73ba8ab941aed36a1e4"} Dec 02 02:12:40 crc kubenswrapper[4884]: I1202 02:12:40.098924 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g45qg" event={"ID":"d8ecb249-9a8d-442a-86ae-d4d28bd3af21","Type":"ContainerStarted","Data":"a0c2e87f47ade142b669536f29dd86b0bfd14f2470a90c35c6d9ced30f72ca15"} Dec 02 02:12:40 crc kubenswrapper[4884]: I1202 02:12:40.102109 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tgs2" event={"ID":"8bf42fe2-0faf-480f-baa0-5a987e80c4f0","Type":"ContainerStarted","Data":"4cd6abe5e7e18813fd355a8f36c75bf2d208c45e08fcf1f7fcc5d89df522770b"} Dec 02 02:12:40 crc kubenswrapper[4884]: I1202 02:12:40.128535 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g45qg" podStartSLOduration=3.023941328 podStartE2EDuration="6.128516646s" podCreationTimestamp="2025-12-02 02:12:34 +0000 UTC" firstStartedPulling="2025-12-02 02:12:36.030325602 +0000 UTC m=+2172.706162506" lastFinishedPulling="2025-12-02 02:12:39.13490093 +0000 UTC m=+2175.810737824" observedRunningTime="2025-12-02 02:12:40.123044132 +0000 UTC m=+2176.798881016" watchObservedRunningTime="2025-12-02 02:12:40.128516646 +0000 UTC m=+2176.804353530" Dec 02 02:12:40 crc kubenswrapper[4884]: I1202 02:12:40.150912 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6tgs2" podStartSLOduration=3.099480674 podStartE2EDuration="6.150886976s" podCreationTimestamp="2025-12-02 02:12:34 +0000 UTC" firstStartedPulling="2025-12-02 02:12:36.034536785 +0000 UTC m=+2172.710373679" lastFinishedPulling="2025-12-02 02:12:39.085943087 +0000 UTC m=+2175.761779981" observedRunningTime="2025-12-02 02:12:40.147873602 +0000 UTC m=+2176.823710486" watchObservedRunningTime="2025-12-02 02:12:40.150886976 +0000 UTC m=+2176.826723870" Dec 02 02:12:44 crc kubenswrapper[4884]: I1202 02:12:44.780601 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:44 crc kubenswrapper[4884]: I1202 02:12:44.781369 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:44 crc kubenswrapper[4884]: I1202 02:12:44.881454 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:44 crc kubenswrapper[4884]: I1202 02:12:44.973073 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:44 crc kubenswrapper[4884]: I1202 02:12:44.973136 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:45 crc kubenswrapper[4884]: I1202 02:12:45.030734 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:45 crc kubenswrapper[4884]: I1202 02:12:45.238994 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:45 crc kubenswrapper[4884]: I1202 02:12:45.242232 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:46 crc kubenswrapper[4884]: I1202 02:12:46.851275 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g45qg"] Dec 02 02:12:46 crc kubenswrapper[4884]: I1202 02:12:46.972535 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:12:46 crc kubenswrapper[4884]: I1202 02:12:46.972639 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:12:46 crc kubenswrapper[4884]: I1202 02:12:46.972826 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 02:12:46 crc kubenswrapper[4884]: I1202 02:12:46.974175 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c0641b565628601f6c7f6d17e6aebb25f493b176f7f04ae6e89252cccd560d8e"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 02:12:46 crc kubenswrapper[4884]: I1202 02:12:46.974288 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://c0641b565628601f6c7f6d17e6aebb25f493b176f7f04ae6e89252cccd560d8e" gracePeriod=600 Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.195014 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="c0641b565628601f6c7f6d17e6aebb25f493b176f7f04ae6e89252cccd560d8e" exitCode=0 Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.195103 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"c0641b565628601f6c7f6d17e6aebb25f493b176f7f04ae6e89252cccd560d8e"} Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.195187 4884 scope.go:117] "RemoveContainer" containerID="e08388be48f367634b8438b8c719502bc679547308aeec33f2cc02a054c404ac" Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.195309 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g45qg" podUID="d8ecb249-9a8d-442a-86ae-d4d28bd3af21" containerName="registry-server" containerID="cri-o://a0c2e87f47ade142b669536f29dd86b0bfd14f2470a90c35c6d9ced30f72ca15" gracePeriod=2 Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.432308 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tgs2"] Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.432851 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6tgs2" podUID="8bf42fe2-0faf-480f-baa0-5a987e80c4f0" containerName="registry-server" containerID="cri-o://4cd6abe5e7e18813fd355a8f36c75bf2d208c45e08fcf1f7fcc5d89df522770b" gracePeriod=2 Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.709165 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.807040 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-utilities\") pod \"d8ecb249-9a8d-442a-86ae-d4d28bd3af21\" (UID: \"d8ecb249-9a8d-442a-86ae-d4d28bd3af21\") " Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.807242 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d77h6\" (UniqueName: \"kubernetes.io/projected/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-kube-api-access-d77h6\") pod \"d8ecb249-9a8d-442a-86ae-d4d28bd3af21\" (UID: \"d8ecb249-9a8d-442a-86ae-d4d28bd3af21\") " Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.807287 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-catalog-content\") pod \"d8ecb249-9a8d-442a-86ae-d4d28bd3af21\" (UID: \"d8ecb249-9a8d-442a-86ae-d4d28bd3af21\") " Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.808294 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-utilities" (OuterVolumeSpecName: "utilities") pod "d8ecb249-9a8d-442a-86ae-d4d28bd3af21" (UID: "d8ecb249-9a8d-442a-86ae-d4d28bd3af21"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.813013 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-kube-api-access-d77h6" (OuterVolumeSpecName: "kube-api-access-d77h6") pod "d8ecb249-9a8d-442a-86ae-d4d28bd3af21" (UID: "d8ecb249-9a8d-442a-86ae-d4d28bd3af21"). InnerVolumeSpecName "kube-api-access-d77h6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.832708 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.879312 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8ecb249-9a8d-442a-86ae-d4d28bd3af21" (UID: "d8ecb249-9a8d-442a-86ae-d4d28bd3af21"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.909090 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.909409 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d77h6\" (UniqueName: \"kubernetes.io/projected/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-kube-api-access-d77h6\") on node \"crc\" DevicePath \"\"" Dec 02 02:12:47 crc kubenswrapper[4884]: I1202 02:12:47.909422 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8ecb249-9a8d-442a-86ae-d4d28bd3af21-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.010907 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-utilities\") pod \"8bf42fe2-0faf-480f-baa0-5a987e80c4f0\" (UID: \"8bf42fe2-0faf-480f-baa0-5a987e80c4f0\") " Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.010961 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-catalog-content\") pod \"8bf42fe2-0faf-480f-baa0-5a987e80c4f0\" (UID: \"8bf42fe2-0faf-480f-baa0-5a987e80c4f0\") " Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.011050 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qlgb\" (UniqueName: \"kubernetes.io/projected/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-kube-api-access-4qlgb\") pod \"8bf42fe2-0faf-480f-baa0-5a987e80c4f0\" (UID: \"8bf42fe2-0faf-480f-baa0-5a987e80c4f0\") " Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.011666 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-utilities" (OuterVolumeSpecName: "utilities") pod "8bf42fe2-0faf-480f-baa0-5a987e80c4f0" (UID: "8bf42fe2-0faf-480f-baa0-5a987e80c4f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.014978 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-kube-api-access-4qlgb" (OuterVolumeSpecName: "kube-api-access-4qlgb") pod "8bf42fe2-0faf-480f-baa0-5a987e80c4f0" (UID: "8bf42fe2-0faf-480f-baa0-5a987e80c4f0"). InnerVolumeSpecName "kube-api-access-4qlgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.028353 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8bf42fe2-0faf-480f-baa0-5a987e80c4f0" (UID: "8bf42fe2-0faf-480f-baa0-5a987e80c4f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.115731 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.115820 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.115845 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qlgb\" (UniqueName: \"kubernetes.io/projected/8bf42fe2-0faf-480f-baa0-5a987e80c4f0-kube-api-access-4qlgb\") on node \"crc\" DevicePath \"\"" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.211388 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653"} Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.215342 4884 generic.go:334] "Generic (PLEG): container finished" podID="8bf42fe2-0faf-480f-baa0-5a987e80c4f0" containerID="4cd6abe5e7e18813fd355a8f36c75bf2d208c45e08fcf1f7fcc5d89df522770b" exitCode=0 Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.215377 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tgs2" event={"ID":"8bf42fe2-0faf-480f-baa0-5a987e80c4f0","Type":"ContainerDied","Data":"4cd6abe5e7e18813fd355a8f36c75bf2d208c45e08fcf1f7fcc5d89df522770b"} Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.215415 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6tgs2" event={"ID":"8bf42fe2-0faf-480f-baa0-5a987e80c4f0","Type":"ContainerDied","Data":"5ac074ee2d19eaa9e8f19f3dedfc5e888eb6378abda26cd434084a40730b1c3e"} Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.215431 4884 scope.go:117] "RemoveContainer" containerID="4cd6abe5e7e18813fd355a8f36c75bf2d208c45e08fcf1f7fcc5d89df522770b" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.215498 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6tgs2" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.218439 4884 generic.go:334] "Generic (PLEG): container finished" podID="d8ecb249-9a8d-442a-86ae-d4d28bd3af21" containerID="a0c2e87f47ade142b669536f29dd86b0bfd14f2470a90c35c6d9ced30f72ca15" exitCode=0 Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.218467 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g45qg" event={"ID":"d8ecb249-9a8d-442a-86ae-d4d28bd3af21","Type":"ContainerDied","Data":"a0c2e87f47ade142b669536f29dd86b0bfd14f2470a90c35c6d9ced30f72ca15"} Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.218486 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g45qg" event={"ID":"d8ecb249-9a8d-442a-86ae-d4d28bd3af21","Type":"ContainerDied","Data":"8047505ac4deb0ba78ea434be0ff9cbfbd57b6b666ad0d77035d897f8e538430"} Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.218575 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g45qg" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.246968 4884 scope.go:117] "RemoveContainer" containerID="384f40eabd18712e702e55e7d8fe0183e8b055078453e73ba8ab941aed36a1e4" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.287720 4884 scope.go:117] "RemoveContainer" containerID="e582571a7febc3149bc9ebeb07e245ea3e3e56ab24ac2064c234a6e07da22144" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.296884 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g45qg"] Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.307307 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g45qg"] Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.316662 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tgs2"] Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.336183 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6tgs2"] Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.338599 4884 scope.go:117] "RemoveContainer" containerID="4cd6abe5e7e18813fd355a8f36c75bf2d208c45e08fcf1f7fcc5d89df522770b" Dec 02 02:12:48 crc kubenswrapper[4884]: E1202 02:12:48.339139 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cd6abe5e7e18813fd355a8f36c75bf2d208c45e08fcf1f7fcc5d89df522770b\": container with ID starting with 4cd6abe5e7e18813fd355a8f36c75bf2d208c45e08fcf1f7fcc5d89df522770b not found: ID does not exist" containerID="4cd6abe5e7e18813fd355a8f36c75bf2d208c45e08fcf1f7fcc5d89df522770b" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.339183 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cd6abe5e7e18813fd355a8f36c75bf2d208c45e08fcf1f7fcc5d89df522770b"} err="failed to get container status \"4cd6abe5e7e18813fd355a8f36c75bf2d208c45e08fcf1f7fcc5d89df522770b\": rpc error: code = NotFound desc = could not find container \"4cd6abe5e7e18813fd355a8f36c75bf2d208c45e08fcf1f7fcc5d89df522770b\": container with ID starting with 4cd6abe5e7e18813fd355a8f36c75bf2d208c45e08fcf1f7fcc5d89df522770b not found: ID does not exist" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.339303 4884 scope.go:117] "RemoveContainer" containerID="384f40eabd18712e702e55e7d8fe0183e8b055078453e73ba8ab941aed36a1e4" Dec 02 02:12:48 crc kubenswrapper[4884]: E1202 02:12:48.339829 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"384f40eabd18712e702e55e7d8fe0183e8b055078453e73ba8ab941aed36a1e4\": container with ID starting with 384f40eabd18712e702e55e7d8fe0183e8b055078453e73ba8ab941aed36a1e4 not found: ID does not exist" containerID="384f40eabd18712e702e55e7d8fe0183e8b055078453e73ba8ab941aed36a1e4" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.339867 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"384f40eabd18712e702e55e7d8fe0183e8b055078453e73ba8ab941aed36a1e4"} err="failed to get container status \"384f40eabd18712e702e55e7d8fe0183e8b055078453e73ba8ab941aed36a1e4\": rpc error: code = NotFound desc = could not find container \"384f40eabd18712e702e55e7d8fe0183e8b055078453e73ba8ab941aed36a1e4\": container with ID starting with 384f40eabd18712e702e55e7d8fe0183e8b055078453e73ba8ab941aed36a1e4 not found: ID does not exist" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.339894 4884 scope.go:117] "RemoveContainer" containerID="e582571a7febc3149bc9ebeb07e245ea3e3e56ab24ac2064c234a6e07da22144" Dec 02 02:12:48 crc kubenswrapper[4884]: E1202 02:12:48.340133 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e582571a7febc3149bc9ebeb07e245ea3e3e56ab24ac2064c234a6e07da22144\": container with ID starting with e582571a7febc3149bc9ebeb07e245ea3e3e56ab24ac2064c234a6e07da22144 not found: ID does not exist" containerID="e582571a7febc3149bc9ebeb07e245ea3e3e56ab24ac2064c234a6e07da22144" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.340169 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e582571a7febc3149bc9ebeb07e245ea3e3e56ab24ac2064c234a6e07da22144"} err="failed to get container status \"e582571a7febc3149bc9ebeb07e245ea3e3e56ab24ac2064c234a6e07da22144\": rpc error: code = NotFound desc = could not find container \"e582571a7febc3149bc9ebeb07e245ea3e3e56ab24ac2064c234a6e07da22144\": container with ID starting with e582571a7febc3149bc9ebeb07e245ea3e3e56ab24ac2064c234a6e07da22144 not found: ID does not exist" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.340191 4884 scope.go:117] "RemoveContainer" containerID="a0c2e87f47ade142b669536f29dd86b0bfd14f2470a90c35c6d9ced30f72ca15" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.375962 4884 scope.go:117] "RemoveContainer" containerID="82ba5c1a84c691497c8ed8d1be3caed480647326b4ee53dbd1de911f5787de26" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.413659 4884 scope.go:117] "RemoveContainer" containerID="e883433eb2bfaf2913ed00212542355e93def7a0cbde5d527d0c3ed54f3f2cd5" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.447904 4884 scope.go:117] "RemoveContainer" containerID="a0c2e87f47ade142b669536f29dd86b0bfd14f2470a90c35c6d9ced30f72ca15" Dec 02 02:12:48 crc kubenswrapper[4884]: E1202 02:12:48.448383 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0c2e87f47ade142b669536f29dd86b0bfd14f2470a90c35c6d9ced30f72ca15\": container with ID starting with a0c2e87f47ade142b669536f29dd86b0bfd14f2470a90c35c6d9ced30f72ca15 not found: ID does not exist" containerID="a0c2e87f47ade142b669536f29dd86b0bfd14f2470a90c35c6d9ced30f72ca15" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.448420 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0c2e87f47ade142b669536f29dd86b0bfd14f2470a90c35c6d9ced30f72ca15"} err="failed to get container status \"a0c2e87f47ade142b669536f29dd86b0bfd14f2470a90c35c6d9ced30f72ca15\": rpc error: code = NotFound desc = could not find container \"a0c2e87f47ade142b669536f29dd86b0bfd14f2470a90c35c6d9ced30f72ca15\": container with ID starting with a0c2e87f47ade142b669536f29dd86b0bfd14f2470a90c35c6d9ced30f72ca15 not found: ID does not exist" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.448445 4884 scope.go:117] "RemoveContainer" containerID="82ba5c1a84c691497c8ed8d1be3caed480647326b4ee53dbd1de911f5787de26" Dec 02 02:12:48 crc kubenswrapper[4884]: E1202 02:12:48.449192 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82ba5c1a84c691497c8ed8d1be3caed480647326b4ee53dbd1de911f5787de26\": container with ID starting with 82ba5c1a84c691497c8ed8d1be3caed480647326b4ee53dbd1de911f5787de26 not found: ID does not exist" containerID="82ba5c1a84c691497c8ed8d1be3caed480647326b4ee53dbd1de911f5787de26" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.449216 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82ba5c1a84c691497c8ed8d1be3caed480647326b4ee53dbd1de911f5787de26"} err="failed to get container status \"82ba5c1a84c691497c8ed8d1be3caed480647326b4ee53dbd1de911f5787de26\": rpc error: code = NotFound desc = could not find container \"82ba5c1a84c691497c8ed8d1be3caed480647326b4ee53dbd1de911f5787de26\": container with ID starting with 82ba5c1a84c691497c8ed8d1be3caed480647326b4ee53dbd1de911f5787de26 not found: ID does not exist" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.449231 4884 scope.go:117] "RemoveContainer" containerID="e883433eb2bfaf2913ed00212542355e93def7a0cbde5d527d0c3ed54f3f2cd5" Dec 02 02:12:48 crc kubenswrapper[4884]: E1202 02:12:48.449530 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e883433eb2bfaf2913ed00212542355e93def7a0cbde5d527d0c3ed54f3f2cd5\": container with ID starting with e883433eb2bfaf2913ed00212542355e93def7a0cbde5d527d0c3ed54f3f2cd5 not found: ID does not exist" containerID="e883433eb2bfaf2913ed00212542355e93def7a0cbde5d527d0c3ed54f3f2cd5" Dec 02 02:12:48 crc kubenswrapper[4884]: I1202 02:12:48.449570 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e883433eb2bfaf2913ed00212542355e93def7a0cbde5d527d0c3ed54f3f2cd5"} err="failed to get container status \"e883433eb2bfaf2913ed00212542355e93def7a0cbde5d527d0c3ed54f3f2cd5\": rpc error: code = NotFound desc = could not find container \"e883433eb2bfaf2913ed00212542355e93def7a0cbde5d527d0c3ed54f3f2cd5\": container with ID starting with e883433eb2bfaf2913ed00212542355e93def7a0cbde5d527d0c3ed54f3f2cd5 not found: ID does not exist" Dec 02 02:12:49 crc kubenswrapper[4884]: I1202 02:12:49.626331 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bf42fe2-0faf-480f-baa0-5a987e80c4f0" path="/var/lib/kubelet/pods/8bf42fe2-0faf-480f-baa0-5a987e80c4f0/volumes" Dec 02 02:12:49 crc kubenswrapper[4884]: I1202 02:12:49.627494 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8ecb249-9a8d-442a-86ae-d4d28bd3af21" path="/var/lib/kubelet/pods/d8ecb249-9a8d-442a-86ae-d4d28bd3af21/volumes" Dec 02 02:12:58 crc kubenswrapper[4884]: I1202 02:12:58.340950 4884 generic.go:334] "Generic (PLEG): container finished" podID="c129a2ce-c6ab-47b5-af8f-55dea1e97d90" containerID="877b9757443a6983258b6ec383d20845859e8874ee48bb48dd4794b2b9ec21e8" exitCode=0 Dec 02 02:12:58 crc kubenswrapper[4884]: I1202 02:12:58.341165 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" event={"ID":"c129a2ce-c6ab-47b5-af8f-55dea1e97d90","Type":"ContainerDied","Data":"877b9757443a6983258b6ec383d20845859e8874ee48bb48dd4794b2b9ec21e8"} Dec 02 02:12:59 crc kubenswrapper[4884]: I1202 02:12:59.884256 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" Dec 02 02:12:59 crc kubenswrapper[4884]: I1202 02:12:59.990310 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-inventory\") pod \"c129a2ce-c6ab-47b5-af8f-55dea1e97d90\" (UID: \"c129a2ce-c6ab-47b5-af8f-55dea1e97d90\") " Dec 02 02:12:59 crc kubenswrapper[4884]: I1202 02:12:59.990374 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-ssh-key\") pod \"c129a2ce-c6ab-47b5-af8f-55dea1e97d90\" (UID: \"c129a2ce-c6ab-47b5-af8f-55dea1e97d90\") " Dec 02 02:12:59 crc kubenswrapper[4884]: I1202 02:12:59.990404 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pss74\" (UniqueName: \"kubernetes.io/projected/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-kube-api-access-pss74\") pod \"c129a2ce-c6ab-47b5-af8f-55dea1e97d90\" (UID: \"c129a2ce-c6ab-47b5-af8f-55dea1e97d90\") " Dec 02 02:12:59 crc kubenswrapper[4884]: I1202 02:12:59.996361 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-kube-api-access-pss74" (OuterVolumeSpecName: "kube-api-access-pss74") pod "c129a2ce-c6ab-47b5-af8f-55dea1e97d90" (UID: "c129a2ce-c6ab-47b5-af8f-55dea1e97d90"). InnerVolumeSpecName "kube-api-access-pss74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.028583 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-inventory" (OuterVolumeSpecName: "inventory") pod "c129a2ce-c6ab-47b5-af8f-55dea1e97d90" (UID: "c129a2ce-c6ab-47b5-af8f-55dea1e97d90"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.036263 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c129a2ce-c6ab-47b5-af8f-55dea1e97d90" (UID: "c129a2ce-c6ab-47b5-af8f-55dea1e97d90"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.093708 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.093797 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.093819 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pss74\" (UniqueName: \"kubernetes.io/projected/c129a2ce-c6ab-47b5-af8f-55dea1e97d90-kube-api-access-pss74\") on node \"crc\" DevicePath \"\"" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.371701 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" event={"ID":"c129a2ce-c6ab-47b5-af8f-55dea1e97d90","Type":"ContainerDied","Data":"23bebc80e2d3d7eeecb9e83470566463cd73430e99310346f329c69d7d8777c0"} Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.371806 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4gns2" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.371823 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23bebc80e2d3d7eeecb9e83470566463cd73430e99310346f329c69d7d8777c0" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.509722 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq"] Dec 02 02:13:00 crc kubenswrapper[4884]: E1202 02:13:00.510231 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c129a2ce-c6ab-47b5-af8f-55dea1e97d90" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.510259 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="c129a2ce-c6ab-47b5-af8f-55dea1e97d90" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 02 02:13:00 crc kubenswrapper[4884]: E1202 02:13:00.510297 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf42fe2-0faf-480f-baa0-5a987e80c4f0" containerName="extract-utilities" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.510307 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf42fe2-0faf-480f-baa0-5a987e80c4f0" containerName="extract-utilities" Dec 02 02:13:00 crc kubenswrapper[4884]: E1202 02:13:00.510333 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8ecb249-9a8d-442a-86ae-d4d28bd3af21" containerName="registry-server" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.510341 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8ecb249-9a8d-442a-86ae-d4d28bd3af21" containerName="registry-server" Dec 02 02:13:00 crc kubenswrapper[4884]: E1202 02:13:00.510364 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8ecb249-9a8d-442a-86ae-d4d28bd3af21" containerName="extract-utilities" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.510372 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8ecb249-9a8d-442a-86ae-d4d28bd3af21" containerName="extract-utilities" Dec 02 02:13:00 crc kubenswrapper[4884]: E1202 02:13:00.510383 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf42fe2-0faf-480f-baa0-5a987e80c4f0" containerName="extract-content" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.510391 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf42fe2-0faf-480f-baa0-5a987e80c4f0" containerName="extract-content" Dec 02 02:13:00 crc kubenswrapper[4884]: E1202 02:13:00.510406 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf42fe2-0faf-480f-baa0-5a987e80c4f0" containerName="registry-server" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.510416 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf42fe2-0faf-480f-baa0-5a987e80c4f0" containerName="registry-server" Dec 02 02:13:00 crc kubenswrapper[4884]: E1202 02:13:00.510436 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8ecb249-9a8d-442a-86ae-d4d28bd3af21" containerName="extract-content" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.510443 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8ecb249-9a8d-442a-86ae-d4d28bd3af21" containerName="extract-content" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.510672 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="c129a2ce-c6ab-47b5-af8f-55dea1e97d90" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.510687 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8ecb249-9a8d-442a-86ae-d4d28bd3af21" containerName="registry-server" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.510728 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf42fe2-0faf-480f-baa0-5a987e80c4f0" containerName="registry-server" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.511640 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.516838 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.517156 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.517347 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.517871 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.539541 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq"] Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.604914 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c70940a8-abc7-4e98-bf89-378b021e5552-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq\" (UID: \"c70940a8-abc7-4e98-bf89-378b021e5552\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.604997 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2skp\" (UniqueName: \"kubernetes.io/projected/c70940a8-abc7-4e98-bf89-378b021e5552-kube-api-access-m2skp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq\" (UID: \"c70940a8-abc7-4e98-bf89-378b021e5552\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.605341 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c70940a8-abc7-4e98-bf89-378b021e5552-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq\" (UID: \"c70940a8-abc7-4e98-bf89-378b021e5552\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.707616 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2skp\" (UniqueName: \"kubernetes.io/projected/c70940a8-abc7-4e98-bf89-378b021e5552-kube-api-access-m2skp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq\" (UID: \"c70940a8-abc7-4e98-bf89-378b021e5552\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.708586 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c70940a8-abc7-4e98-bf89-378b021e5552-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq\" (UID: \"c70940a8-abc7-4e98-bf89-378b021e5552\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.709094 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c70940a8-abc7-4e98-bf89-378b021e5552-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq\" (UID: \"c70940a8-abc7-4e98-bf89-378b021e5552\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.721903 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c70940a8-abc7-4e98-bf89-378b021e5552-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq\" (UID: \"c70940a8-abc7-4e98-bf89-378b021e5552\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.727632 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c70940a8-abc7-4e98-bf89-378b021e5552-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq\" (UID: \"c70940a8-abc7-4e98-bf89-378b021e5552\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.728906 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2skp\" (UniqueName: \"kubernetes.io/projected/c70940a8-abc7-4e98-bf89-378b021e5552-kube-api-access-m2skp\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq\" (UID: \"c70940a8-abc7-4e98-bf89-378b021e5552\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" Dec 02 02:13:00 crc kubenswrapper[4884]: I1202 02:13:00.834171 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" Dec 02 02:13:01 crc kubenswrapper[4884]: I1202 02:13:01.430294 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq"] Dec 02 02:13:01 crc kubenswrapper[4884]: W1202 02:13:01.438820 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc70940a8_abc7_4e98_bf89_378b021e5552.slice/crio-8d10b27eccd5e9bd7a1e2828e4a57257e640c7d229a7894e8e093618224d40da WatchSource:0}: Error finding container 8d10b27eccd5e9bd7a1e2828e4a57257e640c7d229a7894e8e093618224d40da: Status 404 returned error can't find the container with id 8d10b27eccd5e9bd7a1e2828e4a57257e640c7d229a7894e8e093618224d40da Dec 02 02:13:02 crc kubenswrapper[4884]: I1202 02:13:02.409195 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" event={"ID":"c70940a8-abc7-4e98-bf89-378b021e5552","Type":"ContainerStarted","Data":"431bb91ccfb98f4f870a941c129e0d55c555c1ec9e27eac314eed6f8ac4de6d4"} Dec 02 02:13:02 crc kubenswrapper[4884]: I1202 02:13:02.410128 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" event={"ID":"c70940a8-abc7-4e98-bf89-378b021e5552","Type":"ContainerStarted","Data":"8d10b27eccd5e9bd7a1e2828e4a57257e640c7d229a7894e8e093618224d40da"} Dec 02 02:13:02 crc kubenswrapper[4884]: I1202 02:13:02.438481 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" podStartSLOduration=1.80051879 podStartE2EDuration="2.438457161s" podCreationTimestamp="2025-12-02 02:13:00 +0000 UTC" firstStartedPulling="2025-12-02 02:13:01.441935515 +0000 UTC m=+2198.117772399" lastFinishedPulling="2025-12-02 02:13:02.079873856 +0000 UTC m=+2198.755710770" observedRunningTime="2025-12-02 02:13:02.430944746 +0000 UTC m=+2199.106781650" watchObservedRunningTime="2025-12-02 02:13:02.438457161 +0000 UTC m=+2199.114294055" Dec 02 02:13:08 crc kubenswrapper[4884]: I1202 02:13:08.477894 4884 generic.go:334] "Generic (PLEG): container finished" podID="c70940a8-abc7-4e98-bf89-378b021e5552" containerID="431bb91ccfb98f4f870a941c129e0d55c555c1ec9e27eac314eed6f8ac4de6d4" exitCode=0 Dec 02 02:13:08 crc kubenswrapper[4884]: I1202 02:13:08.478019 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" event={"ID":"c70940a8-abc7-4e98-bf89-378b021e5552","Type":"ContainerDied","Data":"431bb91ccfb98f4f870a941c129e0d55c555c1ec9e27eac314eed6f8ac4de6d4"} Dec 02 02:13:09 crc kubenswrapper[4884]: I1202 02:13:09.936283 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.022737 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c70940a8-abc7-4e98-bf89-378b021e5552-ssh-key\") pod \"c70940a8-abc7-4e98-bf89-378b021e5552\" (UID: \"c70940a8-abc7-4e98-bf89-378b021e5552\") " Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.022812 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2skp\" (UniqueName: \"kubernetes.io/projected/c70940a8-abc7-4e98-bf89-378b021e5552-kube-api-access-m2skp\") pod \"c70940a8-abc7-4e98-bf89-378b021e5552\" (UID: \"c70940a8-abc7-4e98-bf89-378b021e5552\") " Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.022993 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c70940a8-abc7-4e98-bf89-378b021e5552-inventory\") pod \"c70940a8-abc7-4e98-bf89-378b021e5552\" (UID: \"c70940a8-abc7-4e98-bf89-378b021e5552\") " Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.027987 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c70940a8-abc7-4e98-bf89-378b021e5552-kube-api-access-m2skp" (OuterVolumeSpecName: "kube-api-access-m2skp") pod "c70940a8-abc7-4e98-bf89-378b021e5552" (UID: "c70940a8-abc7-4e98-bf89-378b021e5552"). InnerVolumeSpecName "kube-api-access-m2skp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.053972 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c70940a8-abc7-4e98-bf89-378b021e5552-inventory" (OuterVolumeSpecName: "inventory") pod "c70940a8-abc7-4e98-bf89-378b021e5552" (UID: "c70940a8-abc7-4e98-bf89-378b021e5552"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.060225 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c70940a8-abc7-4e98-bf89-378b021e5552-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c70940a8-abc7-4e98-bf89-378b021e5552" (UID: "c70940a8-abc7-4e98-bf89-378b021e5552"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.125335 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c70940a8-abc7-4e98-bf89-378b021e5552-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.125370 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c70940a8-abc7-4e98-bf89-378b021e5552-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.125380 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2skp\" (UniqueName: \"kubernetes.io/projected/c70940a8-abc7-4e98-bf89-378b021e5552-kube-api-access-m2skp\") on node \"crc\" DevicePath \"\"" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.502709 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" event={"ID":"c70940a8-abc7-4e98-bf89-378b021e5552","Type":"ContainerDied","Data":"8d10b27eccd5e9bd7a1e2828e4a57257e640c7d229a7894e8e093618224d40da"} Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.503044 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d10b27eccd5e9bd7a1e2828e4a57257e640c7d229a7894e8e093618224d40da" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.502797 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.590989 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf"] Dec 02 02:13:10 crc kubenswrapper[4884]: E1202 02:13:10.591460 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c70940a8-abc7-4e98-bf89-378b021e5552" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.591479 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="c70940a8-abc7-4e98-bf89-378b021e5552" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.591808 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="c70940a8-abc7-4e98-bf89-378b021e5552" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.592543 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.604776 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.605197 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.605554 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.606045 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.617100 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf"] Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.738501 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/860075b8-af90-4ffa-90aa-1c13451eb458-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mqrmf\" (UID: \"860075b8-af90-4ffa-90aa-1c13451eb458\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.739003 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzq58\" (UniqueName: \"kubernetes.io/projected/860075b8-af90-4ffa-90aa-1c13451eb458-kube-api-access-zzq58\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mqrmf\" (UID: \"860075b8-af90-4ffa-90aa-1c13451eb458\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.739077 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/860075b8-af90-4ffa-90aa-1c13451eb458-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mqrmf\" (UID: \"860075b8-af90-4ffa-90aa-1c13451eb458\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.841427 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzq58\" (UniqueName: \"kubernetes.io/projected/860075b8-af90-4ffa-90aa-1c13451eb458-kube-api-access-zzq58\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mqrmf\" (UID: \"860075b8-af90-4ffa-90aa-1c13451eb458\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.841664 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/860075b8-af90-4ffa-90aa-1c13451eb458-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mqrmf\" (UID: \"860075b8-af90-4ffa-90aa-1c13451eb458\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.841723 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/860075b8-af90-4ffa-90aa-1c13451eb458-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mqrmf\" (UID: \"860075b8-af90-4ffa-90aa-1c13451eb458\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.849844 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/860075b8-af90-4ffa-90aa-1c13451eb458-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mqrmf\" (UID: \"860075b8-af90-4ffa-90aa-1c13451eb458\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.858366 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/860075b8-af90-4ffa-90aa-1c13451eb458-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mqrmf\" (UID: \"860075b8-af90-4ffa-90aa-1c13451eb458\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.862517 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzq58\" (UniqueName: \"kubernetes.io/projected/860075b8-af90-4ffa-90aa-1c13451eb458-kube-api-access-zzq58\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-mqrmf\" (UID: \"860075b8-af90-4ffa-90aa-1c13451eb458\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" Dec 02 02:13:10 crc kubenswrapper[4884]: I1202 02:13:10.910901 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" Dec 02 02:13:11 crc kubenswrapper[4884]: I1202 02:13:11.445912 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf"] Dec 02 02:13:11 crc kubenswrapper[4884]: I1202 02:13:11.519691 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" event={"ID":"860075b8-af90-4ffa-90aa-1c13451eb458","Type":"ContainerStarted","Data":"336684bbd884a7ca2bd4a4684223e0fed44f855f533888dad68653cf8fe65e98"} Dec 02 02:13:12 crc kubenswrapper[4884]: I1202 02:13:12.536480 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" event={"ID":"860075b8-af90-4ffa-90aa-1c13451eb458","Type":"ContainerStarted","Data":"f0316d1bc903c7e5066fa67bebca155573dbee39d1ed546f1f0d1ffcfb128e84"} Dec 02 02:13:12 crc kubenswrapper[4884]: I1202 02:13:12.566190 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" podStartSLOduration=2.140468213 podStartE2EDuration="2.566162092s" podCreationTimestamp="2025-12-02 02:13:10 +0000 UTC" firstStartedPulling="2025-12-02 02:13:11.453551018 +0000 UTC m=+2208.129387902" lastFinishedPulling="2025-12-02 02:13:11.879244897 +0000 UTC m=+2208.555081781" observedRunningTime="2025-12-02 02:13:12.554076619 +0000 UTC m=+2209.229913573" watchObservedRunningTime="2025-12-02 02:13:12.566162092 +0000 UTC m=+2209.241999016" Dec 02 02:13:58 crc kubenswrapper[4884]: I1202 02:13:58.033303 4884 generic.go:334] "Generic (PLEG): container finished" podID="860075b8-af90-4ffa-90aa-1c13451eb458" containerID="f0316d1bc903c7e5066fa67bebca155573dbee39d1ed546f1f0d1ffcfb128e84" exitCode=0 Dec 02 02:13:58 crc kubenswrapper[4884]: I1202 02:13:58.033399 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" event={"ID":"860075b8-af90-4ffa-90aa-1c13451eb458","Type":"ContainerDied","Data":"f0316d1bc903c7e5066fa67bebca155573dbee39d1ed546f1f0d1ffcfb128e84"} Dec 02 02:13:59 crc kubenswrapper[4884]: I1202 02:13:59.497419 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" Dec 02 02:13:59 crc kubenswrapper[4884]: I1202 02:13:59.593513 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/860075b8-af90-4ffa-90aa-1c13451eb458-inventory\") pod \"860075b8-af90-4ffa-90aa-1c13451eb458\" (UID: \"860075b8-af90-4ffa-90aa-1c13451eb458\") " Dec 02 02:13:59 crc kubenswrapper[4884]: I1202 02:13:59.594004 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/860075b8-af90-4ffa-90aa-1c13451eb458-ssh-key\") pod \"860075b8-af90-4ffa-90aa-1c13451eb458\" (UID: \"860075b8-af90-4ffa-90aa-1c13451eb458\") " Dec 02 02:13:59 crc kubenswrapper[4884]: I1202 02:13:59.594120 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzq58\" (UniqueName: \"kubernetes.io/projected/860075b8-af90-4ffa-90aa-1c13451eb458-kube-api-access-zzq58\") pod \"860075b8-af90-4ffa-90aa-1c13451eb458\" (UID: \"860075b8-af90-4ffa-90aa-1c13451eb458\") " Dec 02 02:13:59 crc kubenswrapper[4884]: I1202 02:13:59.611128 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/860075b8-af90-4ffa-90aa-1c13451eb458-kube-api-access-zzq58" (OuterVolumeSpecName: "kube-api-access-zzq58") pod "860075b8-af90-4ffa-90aa-1c13451eb458" (UID: "860075b8-af90-4ffa-90aa-1c13451eb458"). InnerVolumeSpecName "kube-api-access-zzq58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:13:59 crc kubenswrapper[4884]: I1202 02:13:59.624601 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/860075b8-af90-4ffa-90aa-1c13451eb458-inventory" (OuterVolumeSpecName: "inventory") pod "860075b8-af90-4ffa-90aa-1c13451eb458" (UID: "860075b8-af90-4ffa-90aa-1c13451eb458"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:13:59 crc kubenswrapper[4884]: I1202 02:13:59.626568 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/860075b8-af90-4ffa-90aa-1c13451eb458-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "860075b8-af90-4ffa-90aa-1c13451eb458" (UID: "860075b8-af90-4ffa-90aa-1c13451eb458"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:13:59 crc kubenswrapper[4884]: I1202 02:13:59.696527 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzq58\" (UniqueName: \"kubernetes.io/projected/860075b8-af90-4ffa-90aa-1c13451eb458-kube-api-access-zzq58\") on node \"crc\" DevicePath \"\"" Dec 02 02:13:59 crc kubenswrapper[4884]: I1202 02:13:59.696556 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/860075b8-af90-4ffa-90aa-1c13451eb458-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:13:59 crc kubenswrapper[4884]: I1202 02:13:59.696564 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/860075b8-af90-4ffa-90aa-1c13451eb458-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.058970 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" event={"ID":"860075b8-af90-4ffa-90aa-1c13451eb458","Type":"ContainerDied","Data":"336684bbd884a7ca2bd4a4684223e0fed44f855f533888dad68653cf8fe65e98"} Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.059010 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="336684bbd884a7ca2bd4a4684223e0fed44f855f533888dad68653cf8fe65e98" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.059050 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-mqrmf" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.158297 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf"] Dec 02 02:14:00 crc kubenswrapper[4884]: E1202 02:14:00.158951 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="860075b8-af90-4ffa-90aa-1c13451eb458" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.158978 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="860075b8-af90-4ffa-90aa-1c13451eb458" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.159248 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="860075b8-af90-4ffa-90aa-1c13451eb458" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.160239 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.162226 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.162841 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.167917 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.168132 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.182440 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf"] Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.208220 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e40740f3-80d8-46da-bfdc-467f13f8be12-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tlflf\" (UID: \"e40740f3-80d8-46da-bfdc-467f13f8be12\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.208363 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsf69\" (UniqueName: \"kubernetes.io/projected/e40740f3-80d8-46da-bfdc-467f13f8be12-kube-api-access-lsf69\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tlflf\" (UID: \"e40740f3-80d8-46da-bfdc-467f13f8be12\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.208484 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e40740f3-80d8-46da-bfdc-467f13f8be12-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tlflf\" (UID: \"e40740f3-80d8-46da-bfdc-467f13f8be12\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.310531 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsf69\" (UniqueName: \"kubernetes.io/projected/e40740f3-80d8-46da-bfdc-467f13f8be12-kube-api-access-lsf69\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tlflf\" (UID: \"e40740f3-80d8-46da-bfdc-467f13f8be12\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.310636 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e40740f3-80d8-46da-bfdc-467f13f8be12-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tlflf\" (UID: \"e40740f3-80d8-46da-bfdc-467f13f8be12\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.310685 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e40740f3-80d8-46da-bfdc-467f13f8be12-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tlflf\" (UID: \"e40740f3-80d8-46da-bfdc-467f13f8be12\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.316009 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e40740f3-80d8-46da-bfdc-467f13f8be12-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tlflf\" (UID: \"e40740f3-80d8-46da-bfdc-467f13f8be12\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.316012 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e40740f3-80d8-46da-bfdc-467f13f8be12-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tlflf\" (UID: \"e40740f3-80d8-46da-bfdc-467f13f8be12\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.328831 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsf69\" (UniqueName: \"kubernetes.io/projected/e40740f3-80d8-46da-bfdc-467f13f8be12-kube-api-access-lsf69\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-tlflf\" (UID: \"e40740f3-80d8-46da-bfdc-467f13f8be12\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" Dec 02 02:14:00 crc kubenswrapper[4884]: I1202 02:14:00.476960 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" Dec 02 02:14:01 crc kubenswrapper[4884]: I1202 02:14:01.030643 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf"] Dec 02 02:14:01 crc kubenswrapper[4884]: I1202 02:14:01.070537 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" event={"ID":"e40740f3-80d8-46da-bfdc-467f13f8be12","Type":"ContainerStarted","Data":"bdbc9631aae3e24bb4b0ff40f8baa297e0983f8221a3ba8bc3123c6e685ec005"} Dec 02 02:14:02 crc kubenswrapper[4884]: I1202 02:14:02.084139 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" event={"ID":"e40740f3-80d8-46da-bfdc-467f13f8be12","Type":"ContainerStarted","Data":"0ca59dd6639e35c8c2b752604e316584733e8c48cff788a7903a3e4c1bed46d6"} Dec 02 02:14:02 crc kubenswrapper[4884]: I1202 02:14:02.109069 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" podStartSLOduration=1.5185575980000001 podStartE2EDuration="2.109052468s" podCreationTimestamp="2025-12-02 02:14:00 +0000 UTC" firstStartedPulling="2025-12-02 02:14:01.043350579 +0000 UTC m=+2257.719187463" lastFinishedPulling="2025-12-02 02:14:01.633845459 +0000 UTC m=+2258.309682333" observedRunningTime="2025-12-02 02:14:02.097798745 +0000 UTC m=+2258.773635639" watchObservedRunningTime="2025-12-02 02:14:02.109052468 +0000 UTC m=+2258.784889352" Dec 02 02:15:00 crc kubenswrapper[4884]: I1202 02:15:00.175887 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v"] Dec 02 02:15:00 crc kubenswrapper[4884]: I1202 02:15:00.178649 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" Dec 02 02:15:00 crc kubenswrapper[4884]: I1202 02:15:00.181895 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 02:15:00 crc kubenswrapper[4884]: I1202 02:15:00.186413 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v"] Dec 02 02:15:00 crc kubenswrapper[4884]: I1202 02:15:00.187320 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 02:15:00 crc kubenswrapper[4884]: I1202 02:15:00.250597 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmz86\" (UniqueName: \"kubernetes.io/projected/14a15ad2-77e7-4021-bab0-0594850ba5bb-kube-api-access-hmz86\") pod \"collect-profiles-29410695-t4d8v\" (UID: \"14a15ad2-77e7-4021-bab0-0594850ba5bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" Dec 02 02:15:00 crc kubenswrapper[4884]: I1202 02:15:00.250659 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/14a15ad2-77e7-4021-bab0-0594850ba5bb-secret-volume\") pod \"collect-profiles-29410695-t4d8v\" (UID: \"14a15ad2-77e7-4021-bab0-0594850ba5bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" Dec 02 02:15:00 crc kubenswrapper[4884]: I1202 02:15:00.250932 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/14a15ad2-77e7-4021-bab0-0594850ba5bb-config-volume\") pod \"collect-profiles-29410695-t4d8v\" (UID: \"14a15ad2-77e7-4021-bab0-0594850ba5bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" Dec 02 02:15:00 crc kubenswrapper[4884]: I1202 02:15:00.352353 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmz86\" (UniqueName: \"kubernetes.io/projected/14a15ad2-77e7-4021-bab0-0594850ba5bb-kube-api-access-hmz86\") pod \"collect-profiles-29410695-t4d8v\" (UID: \"14a15ad2-77e7-4021-bab0-0594850ba5bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" Dec 02 02:15:00 crc kubenswrapper[4884]: I1202 02:15:00.352403 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/14a15ad2-77e7-4021-bab0-0594850ba5bb-secret-volume\") pod \"collect-profiles-29410695-t4d8v\" (UID: \"14a15ad2-77e7-4021-bab0-0594850ba5bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" Dec 02 02:15:00 crc kubenswrapper[4884]: I1202 02:15:00.352482 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/14a15ad2-77e7-4021-bab0-0594850ba5bb-config-volume\") pod \"collect-profiles-29410695-t4d8v\" (UID: \"14a15ad2-77e7-4021-bab0-0594850ba5bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" Dec 02 02:15:00 crc kubenswrapper[4884]: I1202 02:15:00.353549 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/14a15ad2-77e7-4021-bab0-0594850ba5bb-config-volume\") pod \"collect-profiles-29410695-t4d8v\" (UID: \"14a15ad2-77e7-4021-bab0-0594850ba5bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" Dec 02 02:15:00 crc kubenswrapper[4884]: I1202 02:15:00.372683 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/14a15ad2-77e7-4021-bab0-0594850ba5bb-secret-volume\") pod \"collect-profiles-29410695-t4d8v\" (UID: \"14a15ad2-77e7-4021-bab0-0594850ba5bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" Dec 02 02:15:00 crc kubenswrapper[4884]: I1202 02:15:00.381949 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmz86\" (UniqueName: \"kubernetes.io/projected/14a15ad2-77e7-4021-bab0-0594850ba5bb-kube-api-access-hmz86\") pod \"collect-profiles-29410695-t4d8v\" (UID: \"14a15ad2-77e7-4021-bab0-0594850ba5bb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" Dec 02 02:15:00 crc kubenswrapper[4884]: I1202 02:15:00.519616 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" Dec 02 02:15:01 crc kubenswrapper[4884]: I1202 02:15:01.041245 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v"] Dec 02 02:15:01 crc kubenswrapper[4884]: W1202 02:15:01.049425 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14a15ad2_77e7_4021_bab0_0594850ba5bb.slice/crio-a9f415b24a360cf5d4dde86c0a754e8d4065e3d3812e120c1e6c266b34ffc60d WatchSource:0}: Error finding container a9f415b24a360cf5d4dde86c0a754e8d4065e3d3812e120c1e6c266b34ffc60d: Status 404 returned error can't find the container with id a9f415b24a360cf5d4dde86c0a754e8d4065e3d3812e120c1e6c266b34ffc60d Dec 02 02:15:01 crc kubenswrapper[4884]: I1202 02:15:01.904126 4884 generic.go:334] "Generic (PLEG): container finished" podID="14a15ad2-77e7-4021-bab0-0594850ba5bb" containerID="ee9cef5f1904c01a727b9d8fd751eee706e65cce1996424423f4b20d41f39a1f" exitCode=0 Dec 02 02:15:01 crc kubenswrapper[4884]: I1202 02:15:01.904508 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" event={"ID":"14a15ad2-77e7-4021-bab0-0594850ba5bb","Type":"ContainerDied","Data":"ee9cef5f1904c01a727b9d8fd751eee706e65cce1996424423f4b20d41f39a1f"} Dec 02 02:15:01 crc kubenswrapper[4884]: I1202 02:15:01.904546 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" event={"ID":"14a15ad2-77e7-4021-bab0-0594850ba5bb","Type":"ContainerStarted","Data":"a9f415b24a360cf5d4dde86c0a754e8d4065e3d3812e120c1e6c266b34ffc60d"} Dec 02 02:15:03 crc kubenswrapper[4884]: I1202 02:15:03.339114 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" Dec 02 02:15:03 crc kubenswrapper[4884]: I1202 02:15:03.417066 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmz86\" (UniqueName: \"kubernetes.io/projected/14a15ad2-77e7-4021-bab0-0594850ba5bb-kube-api-access-hmz86\") pod \"14a15ad2-77e7-4021-bab0-0594850ba5bb\" (UID: \"14a15ad2-77e7-4021-bab0-0594850ba5bb\") " Dec 02 02:15:03 crc kubenswrapper[4884]: I1202 02:15:03.417174 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/14a15ad2-77e7-4021-bab0-0594850ba5bb-config-volume\") pod \"14a15ad2-77e7-4021-bab0-0594850ba5bb\" (UID: \"14a15ad2-77e7-4021-bab0-0594850ba5bb\") " Dec 02 02:15:03 crc kubenswrapper[4884]: I1202 02:15:03.417325 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/14a15ad2-77e7-4021-bab0-0594850ba5bb-secret-volume\") pod \"14a15ad2-77e7-4021-bab0-0594850ba5bb\" (UID: \"14a15ad2-77e7-4021-bab0-0594850ba5bb\") " Dec 02 02:15:03 crc kubenswrapper[4884]: I1202 02:15:03.419147 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14a15ad2-77e7-4021-bab0-0594850ba5bb-config-volume" (OuterVolumeSpecName: "config-volume") pod "14a15ad2-77e7-4021-bab0-0594850ba5bb" (UID: "14a15ad2-77e7-4021-bab0-0594850ba5bb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:15:03 crc kubenswrapper[4884]: I1202 02:15:03.422483 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14a15ad2-77e7-4021-bab0-0594850ba5bb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "14a15ad2-77e7-4021-bab0-0594850ba5bb" (UID: "14a15ad2-77e7-4021-bab0-0594850ba5bb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:15:03 crc kubenswrapper[4884]: I1202 02:15:03.424547 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14a15ad2-77e7-4021-bab0-0594850ba5bb-kube-api-access-hmz86" (OuterVolumeSpecName: "kube-api-access-hmz86") pod "14a15ad2-77e7-4021-bab0-0594850ba5bb" (UID: "14a15ad2-77e7-4021-bab0-0594850ba5bb"). InnerVolumeSpecName "kube-api-access-hmz86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:15:03 crc kubenswrapper[4884]: I1202 02:15:03.519367 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmz86\" (UniqueName: \"kubernetes.io/projected/14a15ad2-77e7-4021-bab0-0594850ba5bb-kube-api-access-hmz86\") on node \"crc\" DevicePath \"\"" Dec 02 02:15:03 crc kubenswrapper[4884]: I1202 02:15:03.519409 4884 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/14a15ad2-77e7-4021-bab0-0594850ba5bb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 02:15:03 crc kubenswrapper[4884]: I1202 02:15:03.519422 4884 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/14a15ad2-77e7-4021-bab0-0594850ba5bb-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 02:15:03 crc kubenswrapper[4884]: I1202 02:15:03.928605 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" event={"ID":"14a15ad2-77e7-4021-bab0-0594850ba5bb","Type":"ContainerDied","Data":"a9f415b24a360cf5d4dde86c0a754e8d4065e3d3812e120c1e6c266b34ffc60d"} Dec 02 02:15:03 crc kubenswrapper[4884]: I1202 02:15:03.929021 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9f415b24a360cf5d4dde86c0a754e8d4065e3d3812e120c1e6c266b34ffc60d" Dec 02 02:15:03 crc kubenswrapper[4884]: I1202 02:15:03.928690 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v" Dec 02 02:15:04 crc kubenswrapper[4884]: I1202 02:15:04.408364 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4"] Dec 02 02:15:04 crc kubenswrapper[4884]: I1202 02:15:04.416662 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410650-wwcg4"] Dec 02 02:15:05 crc kubenswrapper[4884]: I1202 02:15:05.632186 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5" path="/var/lib/kubelet/pods/83fdc9c6-4f70-4331-b059-1d4b4e3f1ea5/volumes" Dec 02 02:15:06 crc kubenswrapper[4884]: I1202 02:15:06.988653 4884 generic.go:334] "Generic (PLEG): container finished" podID="e40740f3-80d8-46da-bfdc-467f13f8be12" containerID="0ca59dd6639e35c8c2b752604e316584733e8c48cff788a7903a3e4c1bed46d6" exitCode=0 Dec 02 02:15:06 crc kubenswrapper[4884]: I1202 02:15:06.989041 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" event={"ID":"e40740f3-80d8-46da-bfdc-467f13f8be12","Type":"ContainerDied","Data":"0ca59dd6639e35c8c2b752604e316584733e8c48cff788a7903a3e4c1bed46d6"} Dec 02 02:15:08 crc kubenswrapper[4884]: I1202 02:15:08.532097 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" Dec 02 02:15:08 crc kubenswrapper[4884]: I1202 02:15:08.627085 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsf69\" (UniqueName: \"kubernetes.io/projected/e40740f3-80d8-46da-bfdc-467f13f8be12-kube-api-access-lsf69\") pod \"e40740f3-80d8-46da-bfdc-467f13f8be12\" (UID: \"e40740f3-80d8-46da-bfdc-467f13f8be12\") " Dec 02 02:15:08 crc kubenswrapper[4884]: I1202 02:15:08.627414 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e40740f3-80d8-46da-bfdc-467f13f8be12-ssh-key\") pod \"e40740f3-80d8-46da-bfdc-467f13f8be12\" (UID: \"e40740f3-80d8-46da-bfdc-467f13f8be12\") " Dec 02 02:15:08 crc kubenswrapper[4884]: I1202 02:15:08.627521 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e40740f3-80d8-46da-bfdc-467f13f8be12-inventory\") pod \"e40740f3-80d8-46da-bfdc-467f13f8be12\" (UID: \"e40740f3-80d8-46da-bfdc-467f13f8be12\") " Dec 02 02:15:08 crc kubenswrapper[4884]: I1202 02:15:08.636068 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e40740f3-80d8-46da-bfdc-467f13f8be12-kube-api-access-lsf69" (OuterVolumeSpecName: "kube-api-access-lsf69") pod "e40740f3-80d8-46da-bfdc-467f13f8be12" (UID: "e40740f3-80d8-46da-bfdc-467f13f8be12"). InnerVolumeSpecName "kube-api-access-lsf69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:15:08 crc kubenswrapper[4884]: I1202 02:15:08.667894 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e40740f3-80d8-46da-bfdc-467f13f8be12-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e40740f3-80d8-46da-bfdc-467f13f8be12" (UID: "e40740f3-80d8-46da-bfdc-467f13f8be12"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:15:08 crc kubenswrapper[4884]: I1202 02:15:08.679120 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e40740f3-80d8-46da-bfdc-467f13f8be12-inventory" (OuterVolumeSpecName: "inventory") pod "e40740f3-80d8-46da-bfdc-467f13f8be12" (UID: "e40740f3-80d8-46da-bfdc-467f13f8be12"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:15:08 crc kubenswrapper[4884]: I1202 02:15:08.731509 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e40740f3-80d8-46da-bfdc-467f13f8be12-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:15:08 crc kubenswrapper[4884]: I1202 02:15:08.731765 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsf69\" (UniqueName: \"kubernetes.io/projected/e40740f3-80d8-46da-bfdc-467f13f8be12-kube-api-access-lsf69\") on node \"crc\" DevicePath \"\"" Dec 02 02:15:08 crc kubenswrapper[4884]: I1202 02:15:08.731890 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e40740f3-80d8-46da-bfdc-467f13f8be12-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.012783 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" event={"ID":"e40740f3-80d8-46da-bfdc-467f13f8be12","Type":"ContainerDied","Data":"bdbc9631aae3e24bb4b0ff40f8baa297e0983f8221a3ba8bc3123c6e685ec005"} Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.012840 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdbc9631aae3e24bb4b0ff40f8baa297e0983f8221a3ba8bc3123c6e685ec005" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.012851 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-tlflf" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.128651 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h5w4t"] Dec 02 02:15:09 crc kubenswrapper[4884]: E1202 02:15:09.129187 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e40740f3-80d8-46da-bfdc-467f13f8be12" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.129214 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="e40740f3-80d8-46da-bfdc-467f13f8be12" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 02 02:15:09 crc kubenswrapper[4884]: E1202 02:15:09.129249 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a15ad2-77e7-4021-bab0-0594850ba5bb" containerName="collect-profiles" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.129259 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a15ad2-77e7-4021-bab0-0594850ba5bb" containerName="collect-profiles" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.129496 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="14a15ad2-77e7-4021-bab0-0594850ba5bb" containerName="collect-profiles" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.129526 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="e40740f3-80d8-46da-bfdc-467f13f8be12" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.130396 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.133102 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.133316 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.133328 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.135149 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.149122 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h5w4t"] Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.242887 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bedee34a-ff45-4905-9e6d-7408127cbd22-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-h5w4t\" (UID: \"bedee34a-ff45-4905-9e6d-7408127cbd22\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.243035 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bedee34a-ff45-4905-9e6d-7408127cbd22-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-h5w4t\" (UID: \"bedee34a-ff45-4905-9e6d-7408127cbd22\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.243305 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scnll\" (UniqueName: \"kubernetes.io/projected/bedee34a-ff45-4905-9e6d-7408127cbd22-kube-api-access-scnll\") pod \"ssh-known-hosts-edpm-deployment-h5w4t\" (UID: \"bedee34a-ff45-4905-9e6d-7408127cbd22\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.345204 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bedee34a-ff45-4905-9e6d-7408127cbd22-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-h5w4t\" (UID: \"bedee34a-ff45-4905-9e6d-7408127cbd22\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.345447 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bedee34a-ff45-4905-9e6d-7408127cbd22-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-h5w4t\" (UID: \"bedee34a-ff45-4905-9e6d-7408127cbd22\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.345570 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scnll\" (UniqueName: \"kubernetes.io/projected/bedee34a-ff45-4905-9e6d-7408127cbd22-kube-api-access-scnll\") pod \"ssh-known-hosts-edpm-deployment-h5w4t\" (UID: \"bedee34a-ff45-4905-9e6d-7408127cbd22\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.352299 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bedee34a-ff45-4905-9e6d-7408127cbd22-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-h5w4t\" (UID: \"bedee34a-ff45-4905-9e6d-7408127cbd22\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.352943 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bedee34a-ff45-4905-9e6d-7408127cbd22-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-h5w4t\" (UID: \"bedee34a-ff45-4905-9e6d-7408127cbd22\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.374134 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scnll\" (UniqueName: \"kubernetes.io/projected/bedee34a-ff45-4905-9e6d-7408127cbd22-kube-api-access-scnll\") pod \"ssh-known-hosts-edpm-deployment-h5w4t\" (UID: \"bedee34a-ff45-4905-9e6d-7408127cbd22\") " pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" Dec 02 02:15:09 crc kubenswrapper[4884]: I1202 02:15:09.460175 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" Dec 02 02:15:10 crc kubenswrapper[4884]: I1202 02:15:10.113456 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h5w4t"] Dec 02 02:15:10 crc kubenswrapper[4884]: W1202 02:15:10.118086 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbedee34a_ff45_4905_9e6d_7408127cbd22.slice/crio-89ac4dcc3c15d28f1d9dcbca2ddcc11823d73a54a1808ff535eb32d7911b0c33 WatchSource:0}: Error finding container 89ac4dcc3c15d28f1d9dcbca2ddcc11823d73a54a1808ff535eb32d7911b0c33: Status 404 returned error can't find the container with id 89ac4dcc3c15d28f1d9dcbca2ddcc11823d73a54a1808ff535eb32d7911b0c33 Dec 02 02:15:11 crc kubenswrapper[4884]: I1202 02:15:11.042007 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" event={"ID":"bedee34a-ff45-4905-9e6d-7408127cbd22","Type":"ContainerStarted","Data":"89ac4dcc3c15d28f1d9dcbca2ddcc11823d73a54a1808ff535eb32d7911b0c33"} Dec 02 02:15:12 crc kubenswrapper[4884]: I1202 02:15:12.056324 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" event={"ID":"bedee34a-ff45-4905-9e6d-7408127cbd22","Type":"ContainerStarted","Data":"3db0a00108e502cd311049c912ffb4d38e7cffecc68ed70eec3eea085a54d342"} Dec 02 02:15:12 crc kubenswrapper[4884]: I1202 02:15:12.076029 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" podStartSLOduration=2.226859357 podStartE2EDuration="3.075996721s" podCreationTimestamp="2025-12-02 02:15:09 +0000 UTC" firstStartedPulling="2025-12-02 02:15:10.121271151 +0000 UTC m=+2326.797108065" lastFinishedPulling="2025-12-02 02:15:10.970408515 +0000 UTC m=+2327.646245429" observedRunningTime="2025-12-02 02:15:12.072958847 +0000 UTC m=+2328.748795761" watchObservedRunningTime="2025-12-02 02:15:12.075996721 +0000 UTC m=+2328.751833655" Dec 02 02:15:16 crc kubenswrapper[4884]: I1202 02:15:16.972840 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:15:16 crc kubenswrapper[4884]: I1202 02:15:16.973488 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:15:19 crc kubenswrapper[4884]: I1202 02:15:19.146804 4884 generic.go:334] "Generic (PLEG): container finished" podID="bedee34a-ff45-4905-9e6d-7408127cbd22" containerID="3db0a00108e502cd311049c912ffb4d38e7cffecc68ed70eec3eea085a54d342" exitCode=0 Dec 02 02:15:19 crc kubenswrapper[4884]: I1202 02:15:19.146918 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" event={"ID":"bedee34a-ff45-4905-9e6d-7408127cbd22","Type":"ContainerDied","Data":"3db0a00108e502cd311049c912ffb4d38e7cffecc68ed70eec3eea085a54d342"} Dec 02 02:15:20 crc kubenswrapper[4884]: I1202 02:15:20.678056 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" Dec 02 02:15:20 crc kubenswrapper[4884]: I1202 02:15:20.737549 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scnll\" (UniqueName: \"kubernetes.io/projected/bedee34a-ff45-4905-9e6d-7408127cbd22-kube-api-access-scnll\") pod \"bedee34a-ff45-4905-9e6d-7408127cbd22\" (UID: \"bedee34a-ff45-4905-9e6d-7408127cbd22\") " Dec 02 02:15:20 crc kubenswrapper[4884]: I1202 02:15:20.737782 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bedee34a-ff45-4905-9e6d-7408127cbd22-ssh-key-openstack-edpm-ipam\") pod \"bedee34a-ff45-4905-9e6d-7408127cbd22\" (UID: \"bedee34a-ff45-4905-9e6d-7408127cbd22\") " Dec 02 02:15:20 crc kubenswrapper[4884]: I1202 02:15:20.737866 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bedee34a-ff45-4905-9e6d-7408127cbd22-inventory-0\") pod \"bedee34a-ff45-4905-9e6d-7408127cbd22\" (UID: \"bedee34a-ff45-4905-9e6d-7408127cbd22\") " Dec 02 02:15:20 crc kubenswrapper[4884]: I1202 02:15:20.747981 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bedee34a-ff45-4905-9e6d-7408127cbd22-kube-api-access-scnll" (OuterVolumeSpecName: "kube-api-access-scnll") pod "bedee34a-ff45-4905-9e6d-7408127cbd22" (UID: "bedee34a-ff45-4905-9e6d-7408127cbd22"). InnerVolumeSpecName "kube-api-access-scnll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:15:20 crc kubenswrapper[4884]: I1202 02:15:20.775910 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bedee34a-ff45-4905-9e6d-7408127cbd22-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "bedee34a-ff45-4905-9e6d-7408127cbd22" (UID: "bedee34a-ff45-4905-9e6d-7408127cbd22"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:15:20 crc kubenswrapper[4884]: I1202 02:15:20.794760 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bedee34a-ff45-4905-9e6d-7408127cbd22-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "bedee34a-ff45-4905-9e6d-7408127cbd22" (UID: "bedee34a-ff45-4905-9e6d-7408127cbd22"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:15:20 crc kubenswrapper[4884]: I1202 02:15:20.847597 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scnll\" (UniqueName: \"kubernetes.io/projected/bedee34a-ff45-4905-9e6d-7408127cbd22-kube-api-access-scnll\") on node \"crc\" DevicePath \"\"" Dec 02 02:15:20 crc kubenswrapper[4884]: I1202 02:15:20.847899 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bedee34a-ff45-4905-9e6d-7408127cbd22-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 02 02:15:20 crc kubenswrapper[4884]: I1202 02:15:20.847919 4884 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/bedee34a-ff45-4905-9e6d-7408127cbd22-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.177116 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" event={"ID":"bedee34a-ff45-4905-9e6d-7408127cbd22","Type":"ContainerDied","Data":"89ac4dcc3c15d28f1d9dcbca2ddcc11823d73a54a1808ff535eb32d7911b0c33"} Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.177588 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89ac4dcc3c15d28f1d9dcbca2ddcc11823d73a54a1808ff535eb32d7911b0c33" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.177265 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h5w4t" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.268774 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb"] Dec 02 02:15:21 crc kubenswrapper[4884]: E1202 02:15:21.269410 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bedee34a-ff45-4905-9e6d-7408127cbd22" containerName="ssh-known-hosts-edpm-deployment" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.269436 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="bedee34a-ff45-4905-9e6d-7408127cbd22" containerName="ssh-known-hosts-edpm-deployment" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.269830 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="bedee34a-ff45-4905-9e6d-7408127cbd22" containerName="ssh-known-hosts-edpm-deployment" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.271025 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.278439 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb"] Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.302200 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.302525 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.303325 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.303647 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.359053 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d52773c0-4f07-46c6-9774-5fba955697be-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4v9qb\" (UID: \"d52773c0-4f07-46c6-9774-5fba955697be\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.359178 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq9hd\" (UniqueName: \"kubernetes.io/projected/d52773c0-4f07-46c6-9774-5fba955697be-kube-api-access-jq9hd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4v9qb\" (UID: \"d52773c0-4f07-46c6-9774-5fba955697be\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.359254 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d52773c0-4f07-46c6-9774-5fba955697be-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4v9qb\" (UID: \"d52773c0-4f07-46c6-9774-5fba955697be\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.461394 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d52773c0-4f07-46c6-9774-5fba955697be-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4v9qb\" (UID: \"d52773c0-4f07-46c6-9774-5fba955697be\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.461539 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d52773c0-4f07-46c6-9774-5fba955697be-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4v9qb\" (UID: \"d52773c0-4f07-46c6-9774-5fba955697be\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.461773 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq9hd\" (UniqueName: \"kubernetes.io/projected/d52773c0-4f07-46c6-9774-5fba955697be-kube-api-access-jq9hd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4v9qb\" (UID: \"d52773c0-4f07-46c6-9774-5fba955697be\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.468267 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d52773c0-4f07-46c6-9774-5fba955697be-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4v9qb\" (UID: \"d52773c0-4f07-46c6-9774-5fba955697be\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.469205 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d52773c0-4f07-46c6-9774-5fba955697be-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4v9qb\" (UID: \"d52773c0-4f07-46c6-9774-5fba955697be\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.495536 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq9hd\" (UniqueName: \"kubernetes.io/projected/d52773c0-4f07-46c6-9774-5fba955697be-kube-api-access-jq9hd\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-4v9qb\" (UID: \"d52773c0-4f07-46c6-9774-5fba955697be\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" Dec 02 02:15:21 crc kubenswrapper[4884]: I1202 02:15:21.635949 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" Dec 02 02:15:22 crc kubenswrapper[4884]: I1202 02:15:22.033143 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb"] Dec 02 02:15:22 crc kubenswrapper[4884]: W1202 02:15:22.034512 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd52773c0_4f07_46c6_9774_5fba955697be.slice/crio-68bc5fce6c8015fb6687d4c275f0ec517700713832203f2949a7ef1b63ac714a WatchSource:0}: Error finding container 68bc5fce6c8015fb6687d4c275f0ec517700713832203f2949a7ef1b63ac714a: Status 404 returned error can't find the container with id 68bc5fce6c8015fb6687d4c275f0ec517700713832203f2949a7ef1b63ac714a Dec 02 02:15:22 crc kubenswrapper[4884]: I1202 02:15:22.192870 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" event={"ID":"d52773c0-4f07-46c6-9774-5fba955697be","Type":"ContainerStarted","Data":"68bc5fce6c8015fb6687d4c275f0ec517700713832203f2949a7ef1b63ac714a"} Dec 02 02:15:23 crc kubenswrapper[4884]: I1202 02:15:23.213045 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" event={"ID":"d52773c0-4f07-46c6-9774-5fba955697be","Type":"ContainerStarted","Data":"79e380f6db0a6e4a33db93596bc98c972e941f5273ac9804bbccd9adc54c7501"} Dec 02 02:15:23 crc kubenswrapper[4884]: I1202 02:15:23.249043 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" podStartSLOduration=1.765375766 podStartE2EDuration="2.249015879s" podCreationTimestamp="2025-12-02 02:15:21 +0000 UTC" firstStartedPulling="2025-12-02 02:15:22.037510559 +0000 UTC m=+2338.713347483" lastFinishedPulling="2025-12-02 02:15:22.521150672 +0000 UTC m=+2339.196987596" observedRunningTime="2025-12-02 02:15:23.242222785 +0000 UTC m=+2339.918059739" watchObservedRunningTime="2025-12-02 02:15:23.249015879 +0000 UTC m=+2339.924852793" Dec 02 02:15:32 crc kubenswrapper[4884]: I1202 02:15:32.323063 4884 generic.go:334] "Generic (PLEG): container finished" podID="d52773c0-4f07-46c6-9774-5fba955697be" containerID="79e380f6db0a6e4a33db93596bc98c972e941f5273ac9804bbccd9adc54c7501" exitCode=0 Dec 02 02:15:32 crc kubenswrapper[4884]: I1202 02:15:32.323191 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" event={"ID":"d52773c0-4f07-46c6-9774-5fba955697be","Type":"ContainerDied","Data":"79e380f6db0a6e4a33db93596bc98c972e941f5273ac9804bbccd9adc54c7501"} Dec 02 02:15:33 crc kubenswrapper[4884]: I1202 02:15:33.841709 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" Dec 02 02:15:33 crc kubenswrapper[4884]: I1202 02:15:33.944064 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq9hd\" (UniqueName: \"kubernetes.io/projected/d52773c0-4f07-46c6-9774-5fba955697be-kube-api-access-jq9hd\") pod \"d52773c0-4f07-46c6-9774-5fba955697be\" (UID: \"d52773c0-4f07-46c6-9774-5fba955697be\") " Dec 02 02:15:33 crc kubenswrapper[4884]: I1202 02:15:33.944251 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d52773c0-4f07-46c6-9774-5fba955697be-inventory\") pod \"d52773c0-4f07-46c6-9774-5fba955697be\" (UID: \"d52773c0-4f07-46c6-9774-5fba955697be\") " Dec 02 02:15:33 crc kubenswrapper[4884]: I1202 02:15:33.944287 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d52773c0-4f07-46c6-9774-5fba955697be-ssh-key\") pod \"d52773c0-4f07-46c6-9774-5fba955697be\" (UID: \"d52773c0-4f07-46c6-9774-5fba955697be\") " Dec 02 02:15:33 crc kubenswrapper[4884]: I1202 02:15:33.959960 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d52773c0-4f07-46c6-9774-5fba955697be-kube-api-access-jq9hd" (OuterVolumeSpecName: "kube-api-access-jq9hd") pod "d52773c0-4f07-46c6-9774-5fba955697be" (UID: "d52773c0-4f07-46c6-9774-5fba955697be"). InnerVolumeSpecName "kube-api-access-jq9hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:15:33 crc kubenswrapper[4884]: I1202 02:15:33.988617 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d52773c0-4f07-46c6-9774-5fba955697be-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d52773c0-4f07-46c6-9774-5fba955697be" (UID: "d52773c0-4f07-46c6-9774-5fba955697be"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.000461 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d52773c0-4f07-46c6-9774-5fba955697be-inventory" (OuterVolumeSpecName: "inventory") pod "d52773c0-4f07-46c6-9774-5fba955697be" (UID: "d52773c0-4f07-46c6-9774-5fba955697be"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.047466 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq9hd\" (UniqueName: \"kubernetes.io/projected/d52773c0-4f07-46c6-9774-5fba955697be-kube-api-access-jq9hd\") on node \"crc\" DevicePath \"\"" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.048003 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d52773c0-4f07-46c6-9774-5fba955697be-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.048061 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d52773c0-4f07-46c6-9774-5fba955697be-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.350391 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" event={"ID":"d52773c0-4f07-46c6-9774-5fba955697be","Type":"ContainerDied","Data":"68bc5fce6c8015fb6687d4c275f0ec517700713832203f2949a7ef1b63ac714a"} Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.350731 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="68bc5fce6c8015fb6687d4c275f0ec517700713832203f2949a7ef1b63ac714a" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.350540 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-4v9qb" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.477523 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf"] Dec 02 02:15:34 crc kubenswrapper[4884]: E1202 02:15:34.478017 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d52773c0-4f07-46c6-9774-5fba955697be" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.478040 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d52773c0-4f07-46c6-9774-5fba955697be" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.478237 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d52773c0-4f07-46c6-9774-5fba955697be" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.478982 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.484678 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.484695 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.484850 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.485257 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.492672 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf"] Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.564817 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sswdg\" (UniqueName: \"kubernetes.io/projected/d2510824-0c28-4011-a19a-39c0dfb0eafd-kube-api-access-sswdg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-trskf\" (UID: \"d2510824-0c28-4011-a19a-39c0dfb0eafd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.565004 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2510824-0c28-4011-a19a-39c0dfb0eafd-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-trskf\" (UID: \"d2510824-0c28-4011-a19a-39c0dfb0eafd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.565040 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2510824-0c28-4011-a19a-39c0dfb0eafd-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-trskf\" (UID: \"d2510824-0c28-4011-a19a-39c0dfb0eafd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.667131 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2510824-0c28-4011-a19a-39c0dfb0eafd-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-trskf\" (UID: \"d2510824-0c28-4011-a19a-39c0dfb0eafd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.667250 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2510824-0c28-4011-a19a-39c0dfb0eafd-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-trskf\" (UID: \"d2510824-0c28-4011-a19a-39c0dfb0eafd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.667425 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sswdg\" (UniqueName: \"kubernetes.io/projected/d2510824-0c28-4011-a19a-39c0dfb0eafd-kube-api-access-sswdg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-trskf\" (UID: \"d2510824-0c28-4011-a19a-39c0dfb0eafd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.673018 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2510824-0c28-4011-a19a-39c0dfb0eafd-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-trskf\" (UID: \"d2510824-0c28-4011-a19a-39c0dfb0eafd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.674589 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2510824-0c28-4011-a19a-39c0dfb0eafd-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-trskf\" (UID: \"d2510824-0c28-4011-a19a-39c0dfb0eafd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.690938 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sswdg\" (UniqueName: \"kubernetes.io/projected/d2510824-0c28-4011-a19a-39c0dfb0eafd-kube-api-access-sswdg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-trskf\" (UID: \"d2510824-0c28-4011-a19a-39c0dfb0eafd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" Dec 02 02:15:34 crc kubenswrapper[4884]: I1202 02:15:34.813541 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" Dec 02 02:15:35 crc kubenswrapper[4884]: I1202 02:15:35.226727 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf"] Dec 02 02:15:35 crc kubenswrapper[4884]: I1202 02:15:35.359621 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" event={"ID":"d2510824-0c28-4011-a19a-39c0dfb0eafd","Type":"ContainerStarted","Data":"90f5ebd79342b7d32206cbdd410f9b78bf574d35f34861c4d2ecfb67b06d0ea4"} Dec 02 02:15:36 crc kubenswrapper[4884]: I1202 02:15:36.375248 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" event={"ID":"d2510824-0c28-4011-a19a-39c0dfb0eafd","Type":"ContainerStarted","Data":"bc58902a0fd9b2584b8e009cc7055b714a3a82600e0c98a2dac240ddf3d79714"} Dec 02 02:15:36 crc kubenswrapper[4884]: I1202 02:15:36.408028 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" podStartSLOduration=1.950740649 podStartE2EDuration="2.407997462s" podCreationTimestamp="2025-12-02 02:15:34 +0000 UTC" firstStartedPulling="2025-12-02 02:15:35.228980539 +0000 UTC m=+2351.904817423" lastFinishedPulling="2025-12-02 02:15:35.686237352 +0000 UTC m=+2352.362074236" observedRunningTime="2025-12-02 02:15:36.400299216 +0000 UTC m=+2353.076136130" watchObservedRunningTime="2025-12-02 02:15:36.407997462 +0000 UTC m=+2353.083834386" Dec 02 02:15:45 crc kubenswrapper[4884]: I1202 02:15:45.819325 4884 scope.go:117] "RemoveContainer" containerID="1f29e1274403e52b260ee461e939515a522aeb1b9ea0882d6e1b619c53638712" Dec 02 02:15:46 crc kubenswrapper[4884]: I1202 02:15:46.972324 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:15:46 crc kubenswrapper[4884]: I1202 02:15:46.974136 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:15:47 crc kubenswrapper[4884]: I1202 02:15:47.517270 4884 generic.go:334] "Generic (PLEG): container finished" podID="d2510824-0c28-4011-a19a-39c0dfb0eafd" containerID="bc58902a0fd9b2584b8e009cc7055b714a3a82600e0c98a2dac240ddf3d79714" exitCode=0 Dec 02 02:15:47 crc kubenswrapper[4884]: I1202 02:15:47.517355 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" event={"ID":"d2510824-0c28-4011-a19a-39c0dfb0eafd","Type":"ContainerDied","Data":"bc58902a0fd9b2584b8e009cc7055b714a3a82600e0c98a2dac240ddf3d79714"} Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.109593 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.216691 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2510824-0c28-4011-a19a-39c0dfb0eafd-ssh-key\") pod \"d2510824-0c28-4011-a19a-39c0dfb0eafd\" (UID: \"d2510824-0c28-4011-a19a-39c0dfb0eafd\") " Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.216830 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2510824-0c28-4011-a19a-39c0dfb0eafd-inventory\") pod \"d2510824-0c28-4011-a19a-39c0dfb0eafd\" (UID: \"d2510824-0c28-4011-a19a-39c0dfb0eafd\") " Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.217065 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sswdg\" (UniqueName: \"kubernetes.io/projected/d2510824-0c28-4011-a19a-39c0dfb0eafd-kube-api-access-sswdg\") pod \"d2510824-0c28-4011-a19a-39c0dfb0eafd\" (UID: \"d2510824-0c28-4011-a19a-39c0dfb0eafd\") " Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.223009 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2510824-0c28-4011-a19a-39c0dfb0eafd-kube-api-access-sswdg" (OuterVolumeSpecName: "kube-api-access-sswdg") pod "d2510824-0c28-4011-a19a-39c0dfb0eafd" (UID: "d2510824-0c28-4011-a19a-39c0dfb0eafd"). InnerVolumeSpecName "kube-api-access-sswdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.249830 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2510824-0c28-4011-a19a-39c0dfb0eafd-inventory" (OuterVolumeSpecName: "inventory") pod "d2510824-0c28-4011-a19a-39c0dfb0eafd" (UID: "d2510824-0c28-4011-a19a-39c0dfb0eafd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.274674 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2510824-0c28-4011-a19a-39c0dfb0eafd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d2510824-0c28-4011-a19a-39c0dfb0eafd" (UID: "d2510824-0c28-4011-a19a-39c0dfb0eafd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.323336 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d2510824-0c28-4011-a19a-39c0dfb0eafd-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.323389 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d2510824-0c28-4011-a19a-39c0dfb0eafd-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.323411 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sswdg\" (UniqueName: \"kubernetes.io/projected/d2510824-0c28-4011-a19a-39c0dfb0eafd-kube-api-access-sswdg\") on node \"crc\" DevicePath \"\"" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.541376 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" event={"ID":"d2510824-0c28-4011-a19a-39c0dfb0eafd","Type":"ContainerDied","Data":"90f5ebd79342b7d32206cbdd410f9b78bf574d35f34861c4d2ecfb67b06d0ea4"} Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.541950 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90f5ebd79342b7d32206cbdd410f9b78bf574d35f34861c4d2ecfb67b06d0ea4" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.541569 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-trskf" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.643804 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt"] Dec 02 02:15:49 crc kubenswrapper[4884]: E1202 02:15:49.644560 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2510824-0c28-4011-a19a-39c0dfb0eafd" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.644579 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2510824-0c28-4011-a19a-39c0dfb0eafd" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.644782 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2510824-0c28-4011-a19a-39c0dfb0eafd" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.647550 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.650462 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.651079 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.651470 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.651903 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.652105 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.652436 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.652716 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.653455 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.658366 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt"] Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.731030 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.731115 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.731159 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.731410 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.731467 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.731547 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.731603 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.731707 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.731795 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.731846 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.731883 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.731927 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.731988 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7f4v\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-kube-api-access-c7f4v\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.732034 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.848333 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.848416 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.848450 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.848479 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.848499 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.848530 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7f4v\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-kube-api-access-c7f4v\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.848561 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.848602 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.848628 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.848649 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.848823 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.848854 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.848902 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.848927 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.854034 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.857232 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.857662 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.859546 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.859683 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.859921 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.863224 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.863546 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.863397 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.864091 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.864178 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.864267 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.871427 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.886741 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7f4v\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-kube-api-access-c7f4v\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-trsmt\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:49 crc kubenswrapper[4884]: I1202 02:15:49.968831 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:15:50 crc kubenswrapper[4884]: I1202 02:15:50.594461 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt"] Dec 02 02:15:51 crc kubenswrapper[4884]: I1202 02:15:51.567186 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" event={"ID":"32e5e682-f72d-44bd-93c1-0fc27c02f99a","Type":"ContainerStarted","Data":"1c4faaef35a72d8ce6487d76bd7aedc6c75542c6e3c8c36c0a1640e2c5515432"} Dec 02 02:15:51 crc kubenswrapper[4884]: I1202 02:15:51.567891 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" event={"ID":"32e5e682-f72d-44bd-93c1-0fc27c02f99a","Type":"ContainerStarted","Data":"3888e4942d22d2071d6d5a884fc6fd63c4cecd921127d09a123b81ff3991913d"} Dec 02 02:15:51 crc kubenswrapper[4884]: I1202 02:15:51.607068 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" podStartSLOduration=2.095028141 podStartE2EDuration="2.607041411s" podCreationTimestamp="2025-12-02 02:15:49 +0000 UTC" firstStartedPulling="2025-12-02 02:15:50.598579638 +0000 UTC m=+2367.274416522" lastFinishedPulling="2025-12-02 02:15:51.110592908 +0000 UTC m=+2367.786429792" observedRunningTime="2025-12-02 02:15:51.585382867 +0000 UTC m=+2368.261219841" watchObservedRunningTime="2025-12-02 02:15:51.607041411 +0000 UTC m=+2368.282878335" Dec 02 02:16:16 crc kubenswrapper[4884]: I1202 02:16:16.971610 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:16:16 crc kubenswrapper[4884]: I1202 02:16:16.972231 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:16:16 crc kubenswrapper[4884]: I1202 02:16:16.972293 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 02:16:16 crc kubenswrapper[4884]: I1202 02:16:16.973310 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 02:16:16 crc kubenswrapper[4884]: I1202 02:16:16.973407 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" gracePeriod=600 Dec 02 02:16:17 crc kubenswrapper[4884]: E1202 02:16:17.105906 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:16:17 crc kubenswrapper[4884]: I1202 02:16:17.905888 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653"} Dec 02 02:16:17 crc kubenswrapper[4884]: I1202 02:16:17.906316 4884 scope.go:117] "RemoveContainer" containerID="c0641b565628601f6c7f6d17e6aebb25f493b176f7f04ae6e89252cccd560d8e" Dec 02 02:16:17 crc kubenswrapper[4884]: I1202 02:16:17.907348 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:16:17 crc kubenswrapper[4884]: E1202 02:16:17.907858 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:16:17 crc kubenswrapper[4884]: I1202 02:16:17.905839 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" exitCode=0 Dec 02 02:16:29 crc kubenswrapper[4884]: I1202 02:16:29.616135 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:16:29 crc kubenswrapper[4884]: E1202 02:16:29.617407 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:16:37 crc kubenswrapper[4884]: I1202 02:16:37.149445 4884 generic.go:334] "Generic (PLEG): container finished" podID="32e5e682-f72d-44bd-93c1-0fc27c02f99a" containerID="1c4faaef35a72d8ce6487d76bd7aedc6c75542c6e3c8c36c0a1640e2c5515432" exitCode=0 Dec 02 02:16:37 crc kubenswrapper[4884]: I1202 02:16:37.149495 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" event={"ID":"32e5e682-f72d-44bd-93c1-0fc27c02f99a","Type":"ContainerDied","Data":"1c4faaef35a72d8ce6487d76bd7aedc6c75542c6e3c8c36c0a1640e2c5515432"} Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.728429 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.847375 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-ssh-key\") pod \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.847460 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.847563 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-neutron-metadata-combined-ca-bundle\") pod \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.847604 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-libvirt-combined-ca-bundle\") pod \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.847669 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-repo-setup-combined-ca-bundle\") pod \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.847898 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-nova-combined-ca-bundle\") pod \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.847969 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-inventory\") pod \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.848023 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.848062 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7f4v\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-kube-api-access-c7f4v\") pod \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.848111 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-bootstrap-combined-ca-bundle\") pod \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.848274 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-ovn-combined-ca-bundle\") pod \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.848405 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.848459 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-telemetry-combined-ca-bundle\") pod \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.848589 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-ovn-default-certs-0\") pod \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\" (UID: \"32e5e682-f72d-44bd-93c1-0fc27c02f99a\") " Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.854964 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "32e5e682-f72d-44bd-93c1-0fc27c02f99a" (UID: "32e5e682-f72d-44bd-93c1-0fc27c02f99a"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.857950 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "32e5e682-f72d-44bd-93c1-0fc27c02f99a" (UID: "32e5e682-f72d-44bd-93c1-0fc27c02f99a"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.858948 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "32e5e682-f72d-44bd-93c1-0fc27c02f99a" (UID: "32e5e682-f72d-44bd-93c1-0fc27c02f99a"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.859084 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "32e5e682-f72d-44bd-93c1-0fc27c02f99a" (UID: "32e5e682-f72d-44bd-93c1-0fc27c02f99a"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.860466 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "32e5e682-f72d-44bd-93c1-0fc27c02f99a" (UID: "32e5e682-f72d-44bd-93c1-0fc27c02f99a"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.860561 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "32e5e682-f72d-44bd-93c1-0fc27c02f99a" (UID: "32e5e682-f72d-44bd-93c1-0fc27c02f99a"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.861900 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-kube-api-access-c7f4v" (OuterVolumeSpecName: "kube-api-access-c7f4v") pod "32e5e682-f72d-44bd-93c1-0fc27c02f99a" (UID: "32e5e682-f72d-44bd-93c1-0fc27c02f99a"). InnerVolumeSpecName "kube-api-access-c7f4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.862116 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "32e5e682-f72d-44bd-93c1-0fc27c02f99a" (UID: "32e5e682-f72d-44bd-93c1-0fc27c02f99a"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.862122 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "32e5e682-f72d-44bd-93c1-0fc27c02f99a" (UID: "32e5e682-f72d-44bd-93c1-0fc27c02f99a"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.863474 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "32e5e682-f72d-44bd-93c1-0fc27c02f99a" (UID: "32e5e682-f72d-44bd-93c1-0fc27c02f99a"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.864379 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "32e5e682-f72d-44bd-93c1-0fc27c02f99a" (UID: "32e5e682-f72d-44bd-93c1-0fc27c02f99a"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.866611 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "32e5e682-f72d-44bd-93c1-0fc27c02f99a" (UID: "32e5e682-f72d-44bd-93c1-0fc27c02f99a"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.889796 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "32e5e682-f72d-44bd-93c1-0fc27c02f99a" (UID: "32e5e682-f72d-44bd-93c1-0fc27c02f99a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.904375 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-inventory" (OuterVolumeSpecName: "inventory") pod "32e5e682-f72d-44bd-93c1-0fc27c02f99a" (UID: "32e5e682-f72d-44bd-93c1-0fc27c02f99a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.951461 4884 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.951993 4884 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.952167 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.952359 4884 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.952520 4884 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.952646 4884 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.952854 4884 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.953009 4884 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.953155 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.953276 4884 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.953418 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7f4v\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-kube-api-access-c7f4v\") on node \"crc\" DevicePath \"\"" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.953549 4884 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.953691 4884 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32e5e682-f72d-44bd-93c1-0fc27c02f99a-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:16:38 crc kubenswrapper[4884]: I1202 02:16:38.953914 4884 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/32e5e682-f72d-44bd-93c1-0fc27c02f99a-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.177456 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" event={"ID":"32e5e682-f72d-44bd-93c1-0fc27c02f99a","Type":"ContainerDied","Data":"3888e4942d22d2071d6d5a884fc6fd63c4cecd921127d09a123b81ff3991913d"} Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.177520 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3888e4942d22d2071d6d5a884fc6fd63c4cecd921127d09a123b81ff3991913d" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.177614 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-trsmt" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.306553 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq"] Dec 02 02:16:39 crc kubenswrapper[4884]: E1202 02:16:39.307093 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32e5e682-f72d-44bd-93c1-0fc27c02f99a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.307116 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="32e5e682-f72d-44bd-93c1-0fc27c02f99a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.307407 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="32e5e682-f72d-44bd-93c1-0fc27c02f99a" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.308147 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.311018 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.311220 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.311387 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.311547 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.311968 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.322196 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq"] Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.464462 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bd5930b2-0f14-44df-bb7c-7d5e08072485-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8nsq\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.464553 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8nsq\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.464696 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8nsq\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.464870 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8nsq\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.464985 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbl5x\" (UniqueName: \"kubernetes.io/projected/bd5930b2-0f14-44df-bb7c-7d5e08072485-kube-api-access-jbl5x\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8nsq\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.567073 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbl5x\" (UniqueName: \"kubernetes.io/projected/bd5930b2-0f14-44df-bb7c-7d5e08072485-kube-api-access-jbl5x\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8nsq\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.567339 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bd5930b2-0f14-44df-bb7c-7d5e08072485-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8nsq\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.567373 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8nsq\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.567431 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8nsq\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.567462 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8nsq\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.569276 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bd5930b2-0f14-44df-bb7c-7d5e08072485-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8nsq\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.574464 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8nsq\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.575502 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8nsq\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.576963 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8nsq\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.607439 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbl5x\" (UniqueName: \"kubernetes.io/projected/bd5930b2-0f14-44df-bb7c-7d5e08072485-kube-api-access-jbl5x\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8nsq\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:39 crc kubenswrapper[4884]: I1202 02:16:39.641506 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:16:40 crc kubenswrapper[4884]: I1202 02:16:40.263906 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq"] Dec 02 02:16:40 crc kubenswrapper[4884]: I1202 02:16:40.277030 4884 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 02:16:41 crc kubenswrapper[4884]: I1202 02:16:41.230586 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" event={"ID":"bd5930b2-0f14-44df-bb7c-7d5e08072485","Type":"ContainerStarted","Data":"43fdc8172b53c6afd3bf6653e8b6954c9c3e378be0c242fcfd5ee3e6394177a6"} Dec 02 02:16:42 crc kubenswrapper[4884]: I1202 02:16:42.245877 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" event={"ID":"bd5930b2-0f14-44df-bb7c-7d5e08072485","Type":"ContainerStarted","Data":"9e27d6cb22b4f592975649a470815b59985c5c6eee177edc633850f8e738f6f9"} Dec 02 02:16:42 crc kubenswrapper[4884]: I1202 02:16:42.270810 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" podStartSLOduration=2.368107012 podStartE2EDuration="3.270786903s" podCreationTimestamp="2025-12-02 02:16:39 +0000 UTC" firstStartedPulling="2025-12-02 02:16:40.276493855 +0000 UTC m=+2416.952330779" lastFinishedPulling="2025-12-02 02:16:41.179173786 +0000 UTC m=+2417.855010670" observedRunningTime="2025-12-02 02:16:42.266979721 +0000 UTC m=+2418.942816625" watchObservedRunningTime="2025-12-02 02:16:42.270786903 +0000 UTC m=+2418.946623817" Dec 02 02:16:44 crc kubenswrapper[4884]: I1202 02:16:44.663573 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:16:44 crc kubenswrapper[4884]: E1202 02:16:44.664375 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:16:56 crc kubenswrapper[4884]: I1202 02:16:56.614668 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:16:56 crc kubenswrapper[4884]: E1202 02:16:56.615927 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:17:11 crc kubenswrapper[4884]: I1202 02:17:11.614212 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:17:11 crc kubenswrapper[4884]: E1202 02:17:11.615329 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:17:25 crc kubenswrapper[4884]: I1202 02:17:25.669684 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:17:25 crc kubenswrapper[4884]: E1202 02:17:25.670560 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:17:39 crc kubenswrapper[4884]: I1202 02:17:39.614070 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:17:39 crc kubenswrapper[4884]: E1202 02:17:39.614834 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:17:45 crc kubenswrapper[4884]: I1202 02:17:45.006574 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s488q"] Dec 02 02:17:45 crc kubenswrapper[4884]: I1202 02:17:45.009471 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:17:45 crc kubenswrapper[4884]: I1202 02:17:45.036101 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s488q"] Dec 02 02:17:45 crc kubenswrapper[4884]: I1202 02:17:45.101500 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-utilities\") pod \"redhat-operators-s488q\" (UID: \"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62\") " pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:17:45 crc kubenswrapper[4884]: I1202 02:17:45.101826 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw5t6\" (UniqueName: \"kubernetes.io/projected/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-kube-api-access-vw5t6\") pod \"redhat-operators-s488q\" (UID: \"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62\") " pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:17:45 crc kubenswrapper[4884]: I1202 02:17:45.102121 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-catalog-content\") pod \"redhat-operators-s488q\" (UID: \"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62\") " pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:17:45 crc kubenswrapper[4884]: I1202 02:17:45.204158 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-catalog-content\") pod \"redhat-operators-s488q\" (UID: \"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62\") " pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:17:45 crc kubenswrapper[4884]: I1202 02:17:45.204241 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-utilities\") pod \"redhat-operators-s488q\" (UID: \"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62\") " pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:17:45 crc kubenswrapper[4884]: I1202 02:17:45.204307 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw5t6\" (UniqueName: \"kubernetes.io/projected/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-kube-api-access-vw5t6\") pod \"redhat-operators-s488q\" (UID: \"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62\") " pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:17:45 crc kubenswrapper[4884]: I1202 02:17:45.204924 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-catalog-content\") pod \"redhat-operators-s488q\" (UID: \"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62\") " pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:17:45 crc kubenswrapper[4884]: I1202 02:17:45.204981 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-utilities\") pod \"redhat-operators-s488q\" (UID: \"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62\") " pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:17:45 crc kubenswrapper[4884]: I1202 02:17:45.223011 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw5t6\" (UniqueName: \"kubernetes.io/projected/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-kube-api-access-vw5t6\") pod \"redhat-operators-s488q\" (UID: \"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62\") " pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:17:45 crc kubenswrapper[4884]: I1202 02:17:45.332780 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:17:45 crc kubenswrapper[4884]: I1202 02:17:45.815342 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s488q"] Dec 02 02:17:46 crc kubenswrapper[4884]: I1202 02:17:46.426366 4884 generic.go:334] "Generic (PLEG): container finished" podID="ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" containerID="c70a40a8cbff85e71729eab477177d159fae6cc30ba76019340083fcc18879b5" exitCode=0 Dec 02 02:17:46 crc kubenswrapper[4884]: I1202 02:17:46.426439 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s488q" event={"ID":"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62","Type":"ContainerDied","Data":"c70a40a8cbff85e71729eab477177d159fae6cc30ba76019340083fcc18879b5"} Dec 02 02:17:46 crc kubenswrapper[4884]: I1202 02:17:46.426651 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s488q" event={"ID":"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62","Type":"ContainerStarted","Data":"39c2648914523ab9e3118388fea41e7e334792eb2d95fb9ecf677b97bc89454b"} Dec 02 02:17:48 crc kubenswrapper[4884]: I1202 02:17:48.451909 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s488q" event={"ID":"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62","Type":"ContainerStarted","Data":"b5b67553413c62a229acfa556616ab1b4ab1a3bc908ac6d5646cae9ced6aaea4"} Dec 02 02:17:50 crc kubenswrapper[4884]: I1202 02:17:50.477277 4884 generic.go:334] "Generic (PLEG): container finished" podID="ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" containerID="b5b67553413c62a229acfa556616ab1b4ab1a3bc908ac6d5646cae9ced6aaea4" exitCode=0 Dec 02 02:17:50 crc kubenswrapper[4884]: I1202 02:17:50.477988 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s488q" event={"ID":"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62","Type":"ContainerDied","Data":"b5b67553413c62a229acfa556616ab1b4ab1a3bc908ac6d5646cae9ced6aaea4"} Dec 02 02:17:51 crc kubenswrapper[4884]: I1202 02:17:51.502597 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s488q" event={"ID":"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62","Type":"ContainerStarted","Data":"0987dbd018d1957bd66d8cebc9ac55a96bb7d6694186ce6189ac02a6a2b75ddf"} Dec 02 02:17:51 crc kubenswrapper[4884]: I1202 02:17:51.527518 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s488q" podStartSLOduration=2.910369701 podStartE2EDuration="7.527490975s" podCreationTimestamp="2025-12-02 02:17:44 +0000 UTC" firstStartedPulling="2025-12-02 02:17:46.42806871 +0000 UTC m=+2483.103905594" lastFinishedPulling="2025-12-02 02:17:51.045189974 +0000 UTC m=+2487.721026868" observedRunningTime="2025-12-02 02:17:51.522146925 +0000 UTC m=+2488.197983869" watchObservedRunningTime="2025-12-02 02:17:51.527490975 +0000 UTC m=+2488.203327889" Dec 02 02:17:51 crc kubenswrapper[4884]: I1202 02:17:51.614254 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:17:51 crc kubenswrapper[4884]: E1202 02:17:51.614807 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:17:55 crc kubenswrapper[4884]: I1202 02:17:55.333588 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:17:55 crc kubenswrapper[4884]: I1202 02:17:55.334473 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:17:56 crc kubenswrapper[4884]: I1202 02:17:56.419304 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-s488q" podUID="ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" containerName="registry-server" probeResult="failure" output=< Dec 02 02:17:56 crc kubenswrapper[4884]: timeout: failed to connect service ":50051" within 1s Dec 02 02:17:56 crc kubenswrapper[4884]: > Dec 02 02:17:59 crc kubenswrapper[4884]: I1202 02:17:59.590115 4884 generic.go:334] "Generic (PLEG): container finished" podID="bd5930b2-0f14-44df-bb7c-7d5e08072485" containerID="9e27d6cb22b4f592975649a470815b59985c5c6eee177edc633850f8e738f6f9" exitCode=0 Dec 02 02:17:59 crc kubenswrapper[4884]: I1202 02:17:59.590542 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" event={"ID":"bd5930b2-0f14-44df-bb7c-7d5e08072485","Type":"ContainerDied","Data":"9e27d6cb22b4f592975649a470815b59985c5c6eee177edc633850f8e738f6f9"} Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.183872 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.263270 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bd5930b2-0f14-44df-bb7c-7d5e08072485-ovncontroller-config-0\") pod \"bd5930b2-0f14-44df-bb7c-7d5e08072485\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.263362 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbl5x\" (UniqueName: \"kubernetes.io/projected/bd5930b2-0f14-44df-bb7c-7d5e08072485-kube-api-access-jbl5x\") pod \"bd5930b2-0f14-44df-bb7c-7d5e08072485\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.263418 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-ssh-key\") pod \"bd5930b2-0f14-44df-bb7c-7d5e08072485\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.263454 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-ovn-combined-ca-bundle\") pod \"bd5930b2-0f14-44df-bb7c-7d5e08072485\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.263642 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-inventory\") pod \"bd5930b2-0f14-44df-bb7c-7d5e08072485\" (UID: \"bd5930b2-0f14-44df-bb7c-7d5e08072485\") " Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.271425 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "bd5930b2-0f14-44df-bb7c-7d5e08072485" (UID: "bd5930b2-0f14-44df-bb7c-7d5e08072485"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.273450 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd5930b2-0f14-44df-bb7c-7d5e08072485-kube-api-access-jbl5x" (OuterVolumeSpecName: "kube-api-access-jbl5x") pod "bd5930b2-0f14-44df-bb7c-7d5e08072485" (UID: "bd5930b2-0f14-44df-bb7c-7d5e08072485"). InnerVolumeSpecName "kube-api-access-jbl5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.323862 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bd5930b2-0f14-44df-bb7c-7d5e08072485" (UID: "bd5930b2-0f14-44df-bb7c-7d5e08072485"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.328036 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd5930b2-0f14-44df-bb7c-7d5e08072485-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "bd5930b2-0f14-44df-bb7c-7d5e08072485" (UID: "bd5930b2-0f14-44df-bb7c-7d5e08072485"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.345142 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-inventory" (OuterVolumeSpecName: "inventory") pod "bd5930b2-0f14-44df-bb7c-7d5e08072485" (UID: "bd5930b2-0f14-44df-bb7c-7d5e08072485"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.366533 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.366578 4884 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bd5930b2-0f14-44df-bb7c-7d5e08072485-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.366595 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbl5x\" (UniqueName: \"kubernetes.io/projected/bd5930b2-0f14-44df-bb7c-7d5e08072485-kube-api-access-jbl5x\") on node \"crc\" DevicePath \"\"" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.366609 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.366621 4884 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd5930b2-0f14-44df-bb7c-7d5e08072485-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.622400 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.632136 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8nsq" event={"ID":"bd5930b2-0f14-44df-bb7c-7d5e08072485","Type":"ContainerDied","Data":"43fdc8172b53c6afd3bf6653e8b6954c9c3e378be0c242fcfd5ee3e6394177a6"} Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.632181 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43fdc8172b53c6afd3bf6653e8b6954c9c3e378be0c242fcfd5ee3e6394177a6" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.920126 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs"] Dec 02 02:18:01 crc kubenswrapper[4884]: E1202 02:18:01.920608 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd5930b2-0f14-44df-bb7c-7d5e08072485" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.920628 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd5930b2-0f14-44df-bb7c-7d5e08072485" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.920917 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd5930b2-0f14-44df-bb7c-7d5e08072485" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.921673 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.924305 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.924330 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.924606 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.924871 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.925443 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.925675 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.938226 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs"] Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.981128 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvgrg\" (UniqueName: \"kubernetes.io/projected/9703a9cb-2e1a-4e9f-a862-8025a775ac87-kube-api-access-jvgrg\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.981213 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.981262 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.981295 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.981361 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:01 crc kubenswrapper[4884]: I1202 02:18:01.981609 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:02 crc kubenswrapper[4884]: I1202 02:18:02.084170 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:02 crc kubenswrapper[4884]: I1202 02:18:02.084324 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:02 crc kubenswrapper[4884]: I1202 02:18:02.084505 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvgrg\" (UniqueName: \"kubernetes.io/projected/9703a9cb-2e1a-4e9f-a862-8025a775ac87-kube-api-access-jvgrg\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:02 crc kubenswrapper[4884]: I1202 02:18:02.084581 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:02 crc kubenswrapper[4884]: I1202 02:18:02.084648 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:02 crc kubenswrapper[4884]: I1202 02:18:02.084709 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:02 crc kubenswrapper[4884]: I1202 02:18:02.089724 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:02 crc kubenswrapper[4884]: I1202 02:18:02.090187 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:02 crc kubenswrapper[4884]: I1202 02:18:02.090628 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:02 crc kubenswrapper[4884]: I1202 02:18:02.091699 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:02 crc kubenswrapper[4884]: I1202 02:18:02.105790 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvgrg\" (UniqueName: \"kubernetes.io/projected/9703a9cb-2e1a-4e9f-a862-8025a775ac87-kube-api-access-jvgrg\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:02 crc kubenswrapper[4884]: I1202 02:18:02.107405 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:02 crc kubenswrapper[4884]: I1202 02:18:02.244320 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:18:02 crc kubenswrapper[4884]: W1202 02:18:02.861096 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9703a9cb_2e1a_4e9f_a862_8025a775ac87.slice/crio-a455acc9b381fd6a91b7ead256466c1c003659a3f47060c8a7fc228d13c7af9b WatchSource:0}: Error finding container a455acc9b381fd6a91b7ead256466c1c003659a3f47060c8a7fc228d13c7af9b: Status 404 returned error can't find the container with id a455acc9b381fd6a91b7ead256466c1c003659a3f47060c8a7fc228d13c7af9b Dec 02 02:18:02 crc kubenswrapper[4884]: I1202 02:18:02.878317 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs"] Dec 02 02:18:03 crc kubenswrapper[4884]: I1202 02:18:03.662829 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" event={"ID":"9703a9cb-2e1a-4e9f-a862-8025a775ac87","Type":"ContainerStarted","Data":"a455acc9b381fd6a91b7ead256466c1c003659a3f47060c8a7fc228d13c7af9b"} Dec 02 02:18:04 crc kubenswrapper[4884]: I1202 02:18:04.674203 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" event={"ID":"9703a9cb-2e1a-4e9f-a862-8025a775ac87","Type":"ContainerStarted","Data":"5dca0bbd42128bc075372bb86310778acbd5d6565508c218111e07ca2e94b22a"} Dec 02 02:18:04 crc kubenswrapper[4884]: I1202 02:18:04.698313 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" podStartSLOduration=3.1886453 podStartE2EDuration="3.698291978s" podCreationTimestamp="2025-12-02 02:18:01 +0000 UTC" firstStartedPulling="2025-12-02 02:18:02.863646568 +0000 UTC m=+2499.539483452" lastFinishedPulling="2025-12-02 02:18:03.373293256 +0000 UTC m=+2500.049130130" observedRunningTime="2025-12-02 02:18:04.694141687 +0000 UTC m=+2501.369978571" watchObservedRunningTime="2025-12-02 02:18:04.698291978 +0000 UTC m=+2501.374128882" Dec 02 02:18:05 crc kubenswrapper[4884]: I1202 02:18:05.438975 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:18:05 crc kubenswrapper[4884]: I1202 02:18:05.508066 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:18:05 crc kubenswrapper[4884]: I1202 02:18:05.691715 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s488q"] Dec 02 02:18:06 crc kubenswrapper[4884]: I1202 02:18:06.614823 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:18:06 crc kubenswrapper[4884]: E1202 02:18:06.615113 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:18:06 crc kubenswrapper[4884]: I1202 02:18:06.700510 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s488q" podUID="ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" containerName="registry-server" containerID="cri-o://0987dbd018d1957bd66d8cebc9ac55a96bb7d6694186ce6189ac02a6a2b75ddf" gracePeriod=2 Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.194574 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.314507 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw5t6\" (UniqueName: \"kubernetes.io/projected/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-kube-api-access-vw5t6\") pod \"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62\" (UID: \"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62\") " Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.314648 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-catalog-content\") pod \"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62\" (UID: \"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62\") " Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.314701 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-utilities\") pod \"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62\" (UID: \"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62\") " Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.315547 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-utilities" (OuterVolumeSpecName: "utilities") pod "ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" (UID: "ed9502ff-5b9e-46d9-9d7b-5398f9a29f62"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.329342 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-kube-api-access-vw5t6" (OuterVolumeSpecName: "kube-api-access-vw5t6") pod "ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" (UID: "ed9502ff-5b9e-46d9-9d7b-5398f9a29f62"). InnerVolumeSpecName "kube-api-access-vw5t6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.414063 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" (UID: "ed9502ff-5b9e-46d9-9d7b-5398f9a29f62"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.416355 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw5t6\" (UniqueName: \"kubernetes.io/projected/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-kube-api-access-vw5t6\") on node \"crc\" DevicePath \"\"" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.416395 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.416411 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.717703 4884 generic.go:334] "Generic (PLEG): container finished" podID="ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" containerID="0987dbd018d1957bd66d8cebc9ac55a96bb7d6694186ce6189ac02a6a2b75ddf" exitCode=0 Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.717789 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s488q" event={"ID":"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62","Type":"ContainerDied","Data":"0987dbd018d1957bd66d8cebc9ac55a96bb7d6694186ce6189ac02a6a2b75ddf"} Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.717831 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s488q" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.717868 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s488q" event={"ID":"ed9502ff-5b9e-46d9-9d7b-5398f9a29f62","Type":"ContainerDied","Data":"39c2648914523ab9e3118388fea41e7e334792eb2d95fb9ecf677b97bc89454b"} Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.717899 4884 scope.go:117] "RemoveContainer" containerID="0987dbd018d1957bd66d8cebc9ac55a96bb7d6694186ce6189ac02a6a2b75ddf" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.752586 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s488q"] Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.754134 4884 scope.go:117] "RemoveContainer" containerID="b5b67553413c62a229acfa556616ab1b4ab1a3bc908ac6d5646cae9ced6aaea4" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.764996 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s488q"] Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.788905 4884 scope.go:117] "RemoveContainer" containerID="c70a40a8cbff85e71729eab477177d159fae6cc30ba76019340083fcc18879b5" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.828371 4884 scope.go:117] "RemoveContainer" containerID="0987dbd018d1957bd66d8cebc9ac55a96bb7d6694186ce6189ac02a6a2b75ddf" Dec 02 02:18:07 crc kubenswrapper[4884]: E1202 02:18:07.829280 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0987dbd018d1957bd66d8cebc9ac55a96bb7d6694186ce6189ac02a6a2b75ddf\": container with ID starting with 0987dbd018d1957bd66d8cebc9ac55a96bb7d6694186ce6189ac02a6a2b75ddf not found: ID does not exist" containerID="0987dbd018d1957bd66d8cebc9ac55a96bb7d6694186ce6189ac02a6a2b75ddf" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.829363 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0987dbd018d1957bd66d8cebc9ac55a96bb7d6694186ce6189ac02a6a2b75ddf"} err="failed to get container status \"0987dbd018d1957bd66d8cebc9ac55a96bb7d6694186ce6189ac02a6a2b75ddf\": rpc error: code = NotFound desc = could not find container \"0987dbd018d1957bd66d8cebc9ac55a96bb7d6694186ce6189ac02a6a2b75ddf\": container with ID starting with 0987dbd018d1957bd66d8cebc9ac55a96bb7d6694186ce6189ac02a6a2b75ddf not found: ID does not exist" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.829397 4884 scope.go:117] "RemoveContainer" containerID="b5b67553413c62a229acfa556616ab1b4ab1a3bc908ac6d5646cae9ced6aaea4" Dec 02 02:18:07 crc kubenswrapper[4884]: E1202 02:18:07.829799 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5b67553413c62a229acfa556616ab1b4ab1a3bc908ac6d5646cae9ced6aaea4\": container with ID starting with b5b67553413c62a229acfa556616ab1b4ab1a3bc908ac6d5646cae9ced6aaea4 not found: ID does not exist" containerID="b5b67553413c62a229acfa556616ab1b4ab1a3bc908ac6d5646cae9ced6aaea4" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.829858 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5b67553413c62a229acfa556616ab1b4ab1a3bc908ac6d5646cae9ced6aaea4"} err="failed to get container status \"b5b67553413c62a229acfa556616ab1b4ab1a3bc908ac6d5646cae9ced6aaea4\": rpc error: code = NotFound desc = could not find container \"b5b67553413c62a229acfa556616ab1b4ab1a3bc908ac6d5646cae9ced6aaea4\": container with ID starting with b5b67553413c62a229acfa556616ab1b4ab1a3bc908ac6d5646cae9ced6aaea4 not found: ID does not exist" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.829880 4884 scope.go:117] "RemoveContainer" containerID="c70a40a8cbff85e71729eab477177d159fae6cc30ba76019340083fcc18879b5" Dec 02 02:18:07 crc kubenswrapper[4884]: E1202 02:18:07.830250 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c70a40a8cbff85e71729eab477177d159fae6cc30ba76019340083fcc18879b5\": container with ID starting with c70a40a8cbff85e71729eab477177d159fae6cc30ba76019340083fcc18879b5 not found: ID does not exist" containerID="c70a40a8cbff85e71729eab477177d159fae6cc30ba76019340083fcc18879b5" Dec 02 02:18:07 crc kubenswrapper[4884]: I1202 02:18:07.830330 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c70a40a8cbff85e71729eab477177d159fae6cc30ba76019340083fcc18879b5"} err="failed to get container status \"c70a40a8cbff85e71729eab477177d159fae6cc30ba76019340083fcc18879b5\": rpc error: code = NotFound desc = could not find container \"c70a40a8cbff85e71729eab477177d159fae6cc30ba76019340083fcc18879b5\": container with ID starting with c70a40a8cbff85e71729eab477177d159fae6cc30ba76019340083fcc18879b5 not found: ID does not exist" Dec 02 02:18:09 crc kubenswrapper[4884]: I1202 02:18:09.656150 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" path="/var/lib/kubelet/pods/ed9502ff-5b9e-46d9-9d7b-5398f9a29f62/volumes" Dec 02 02:18:17 crc kubenswrapper[4884]: I1202 02:18:17.614367 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:18:17 crc kubenswrapper[4884]: E1202 02:18:17.615306 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:18:31 crc kubenswrapper[4884]: I1202 02:18:31.614941 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:18:31 crc kubenswrapper[4884]: E1202 02:18:31.615791 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:18:45 crc kubenswrapper[4884]: I1202 02:18:45.615366 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:18:45 crc kubenswrapper[4884]: E1202 02:18:45.616538 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:18:59 crc kubenswrapper[4884]: I1202 02:18:59.614897 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:18:59 crc kubenswrapper[4884]: E1202 02:18:59.615693 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:19:02 crc kubenswrapper[4884]: I1202 02:19:02.334843 4884 generic.go:334] "Generic (PLEG): container finished" podID="9703a9cb-2e1a-4e9f-a862-8025a775ac87" containerID="5dca0bbd42128bc075372bb86310778acbd5d6565508c218111e07ca2e94b22a" exitCode=0 Dec 02 02:19:02 crc kubenswrapper[4884]: I1202 02:19:02.334938 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" event={"ID":"9703a9cb-2e1a-4e9f-a862-8025a775ac87","Type":"ContainerDied","Data":"5dca0bbd42128bc075372bb86310778acbd5d6565508c218111e07ca2e94b22a"} Dec 02 02:19:03 crc kubenswrapper[4884]: I1202 02:19:03.907612 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.109717 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvgrg\" (UniqueName: \"kubernetes.io/projected/9703a9cb-2e1a-4e9f-a862-8025a775ac87-kube-api-access-jvgrg\") pod \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.109910 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-neutron-metadata-combined-ca-bundle\") pod \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.110038 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-nova-metadata-neutron-config-0\") pod \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.110181 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-inventory\") pod \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.110303 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-ssh-key\") pod \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.110432 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-neutron-ovn-metadata-agent-neutron-config-0\") pod \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\" (UID: \"9703a9cb-2e1a-4e9f-a862-8025a775ac87\") " Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.117209 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "9703a9cb-2e1a-4e9f-a862-8025a775ac87" (UID: "9703a9cb-2e1a-4e9f-a862-8025a775ac87"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.118163 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9703a9cb-2e1a-4e9f-a862-8025a775ac87-kube-api-access-jvgrg" (OuterVolumeSpecName: "kube-api-access-jvgrg") pod "9703a9cb-2e1a-4e9f-a862-8025a775ac87" (UID: "9703a9cb-2e1a-4e9f-a862-8025a775ac87"). InnerVolumeSpecName "kube-api-access-jvgrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.148684 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "9703a9cb-2e1a-4e9f-a862-8025a775ac87" (UID: "9703a9cb-2e1a-4e9f-a862-8025a775ac87"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.149620 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "9703a9cb-2e1a-4e9f-a862-8025a775ac87" (UID: "9703a9cb-2e1a-4e9f-a862-8025a775ac87"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.161771 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9703a9cb-2e1a-4e9f-a862-8025a775ac87" (UID: "9703a9cb-2e1a-4e9f-a862-8025a775ac87"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.170664 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-inventory" (OuterVolumeSpecName: "inventory") pod "9703a9cb-2e1a-4e9f-a862-8025a775ac87" (UID: "9703a9cb-2e1a-4e9f-a862-8025a775ac87"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.216516 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvgrg\" (UniqueName: \"kubernetes.io/projected/9703a9cb-2e1a-4e9f-a862-8025a775ac87-kube-api-access-jvgrg\") on node \"crc\" DevicePath \"\"" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.216575 4884 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.216595 4884 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.216609 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.216621 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.216632 4884 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9703a9cb-2e1a-4e9f-a862-8025a775ac87-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.359857 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" event={"ID":"9703a9cb-2e1a-4e9f-a862-8025a775ac87","Type":"ContainerDied","Data":"a455acc9b381fd6a91b7ead256466c1c003659a3f47060c8a7fc228d13c7af9b"} Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.359913 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a455acc9b381fd6a91b7ead256466c1c003659a3f47060c8a7fc228d13c7af9b" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.359940 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.479986 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb"] Dec 02 02:19:04 crc kubenswrapper[4884]: E1202 02:19:04.480366 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" containerName="registry-server" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.480380 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" containerName="registry-server" Dec 02 02:19:04 crc kubenswrapper[4884]: E1202 02:19:04.480404 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" containerName="extract-utilities" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.480410 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" containerName="extract-utilities" Dec 02 02:19:04 crc kubenswrapper[4884]: E1202 02:19:04.480418 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" containerName="extract-content" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.480423 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" containerName="extract-content" Dec 02 02:19:04 crc kubenswrapper[4884]: E1202 02:19:04.480453 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9703a9cb-2e1a-4e9f-a862-8025a775ac87" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.480459 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="9703a9cb-2e1a-4e9f-a862-8025a775ac87" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.480620 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="9703a9cb-2e1a-4e9f-a862-8025a775ac87" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.480638 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed9502ff-5b9e-46d9-9d7b-5398f9a29f62" containerName="registry-server" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.481258 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.483714 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.484009 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.484557 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.484572 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.484678 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.506142 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb"] Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.540002 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.540115 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.540435 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.540503 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.540576 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6pr8\" (UniqueName: \"kubernetes.io/projected/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-kube-api-access-p6pr8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.642671 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.642723 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.642761 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6pr8\" (UniqueName: \"kubernetes.io/projected/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-kube-api-access-p6pr8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.642847 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.642900 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.649622 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.649645 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.650315 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.653562 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.662116 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6pr8\" (UniqueName: \"kubernetes.io/projected/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-kube-api-access-p6pr8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:04 crc kubenswrapper[4884]: I1202 02:19:04.837920 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:19:05 crc kubenswrapper[4884]: I1202 02:19:05.429570 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb"] Dec 02 02:19:06 crc kubenswrapper[4884]: I1202 02:19:06.396519 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" event={"ID":"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d","Type":"ContainerStarted","Data":"576bb405ff77b646151ed41939d6aff34d31421f203dd64f897404f70f5437c6"} Dec 02 02:19:06 crc kubenswrapper[4884]: I1202 02:19:06.396923 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" event={"ID":"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d","Type":"ContainerStarted","Data":"af2aeeb20547604a60afc3e34fde38f15596b0742ae588cd8149953a4a954fdc"} Dec 02 02:19:06 crc kubenswrapper[4884]: I1202 02:19:06.431081 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" podStartSLOduration=1.911784296 podStartE2EDuration="2.43105875s" podCreationTimestamp="2025-12-02 02:19:04 +0000 UTC" firstStartedPulling="2025-12-02 02:19:05.437860879 +0000 UTC m=+2562.113697773" lastFinishedPulling="2025-12-02 02:19:05.957135343 +0000 UTC m=+2562.632972227" observedRunningTime="2025-12-02 02:19:06.420087512 +0000 UTC m=+2563.095924436" watchObservedRunningTime="2025-12-02 02:19:06.43105875 +0000 UTC m=+2563.106895644" Dec 02 02:19:13 crc kubenswrapper[4884]: I1202 02:19:13.631295 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:19:13 crc kubenswrapper[4884]: E1202 02:19:13.632232 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:19:27 crc kubenswrapper[4884]: I1202 02:19:27.615296 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:19:27 crc kubenswrapper[4884]: E1202 02:19:27.616367 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:19:38 crc kubenswrapper[4884]: I1202 02:19:38.614819 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:19:38 crc kubenswrapper[4884]: E1202 02:19:38.616273 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:19:52 crc kubenswrapper[4884]: I1202 02:19:52.616778 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:19:52 crc kubenswrapper[4884]: E1202 02:19:52.617631 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:20:07 crc kubenswrapper[4884]: I1202 02:20:07.614263 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:20:07 crc kubenswrapper[4884]: E1202 02:20:07.615194 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:20:22 crc kubenswrapper[4884]: I1202 02:20:22.615627 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:20:22 crc kubenswrapper[4884]: E1202 02:20:22.616902 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:20:34 crc kubenswrapper[4884]: I1202 02:20:34.614390 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:20:34 crc kubenswrapper[4884]: E1202 02:20:34.615476 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:20:49 crc kubenswrapper[4884]: I1202 02:20:49.614819 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:20:49 crc kubenswrapper[4884]: E1202 02:20:49.615542 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:21:04 crc kubenswrapper[4884]: I1202 02:21:04.616526 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:21:04 crc kubenswrapper[4884]: E1202 02:21:04.617524 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:21:17 crc kubenswrapper[4884]: I1202 02:21:17.614975 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:21:19 crc kubenswrapper[4884]: I1202 02:21:19.120873 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"83598bb7142be85f0e3373487befb0fe500a455b9b41c9b4af785be9d73c5ae8"} Dec 02 02:22:57 crc kubenswrapper[4884]: I1202 02:22:57.819310 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8gk4c"] Dec 02 02:22:57 crc kubenswrapper[4884]: I1202 02:22:57.845567 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:22:57 crc kubenswrapper[4884]: I1202 02:22:57.869391 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8gk4c"] Dec 02 02:22:58 crc kubenswrapper[4884]: I1202 02:22:58.010033 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84531916-5cd5-425c-b024-a657aa80a781-catalog-content\") pod \"community-operators-8gk4c\" (UID: \"84531916-5cd5-425c-b024-a657aa80a781\") " pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:22:58 crc kubenswrapper[4884]: I1202 02:22:58.010184 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84531916-5cd5-425c-b024-a657aa80a781-utilities\") pod \"community-operators-8gk4c\" (UID: \"84531916-5cd5-425c-b024-a657aa80a781\") " pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:22:58 crc kubenswrapper[4884]: I1202 02:22:58.010227 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhwlf\" (UniqueName: \"kubernetes.io/projected/84531916-5cd5-425c-b024-a657aa80a781-kube-api-access-nhwlf\") pod \"community-operators-8gk4c\" (UID: \"84531916-5cd5-425c-b024-a657aa80a781\") " pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:22:58 crc kubenswrapper[4884]: I1202 02:22:58.111507 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84531916-5cd5-425c-b024-a657aa80a781-utilities\") pod \"community-operators-8gk4c\" (UID: \"84531916-5cd5-425c-b024-a657aa80a781\") " pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:22:58 crc kubenswrapper[4884]: I1202 02:22:58.111911 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhwlf\" (UniqueName: \"kubernetes.io/projected/84531916-5cd5-425c-b024-a657aa80a781-kube-api-access-nhwlf\") pod \"community-operators-8gk4c\" (UID: \"84531916-5cd5-425c-b024-a657aa80a781\") " pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:22:58 crc kubenswrapper[4884]: I1202 02:22:58.112034 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84531916-5cd5-425c-b024-a657aa80a781-utilities\") pod \"community-operators-8gk4c\" (UID: \"84531916-5cd5-425c-b024-a657aa80a781\") " pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:22:58 crc kubenswrapper[4884]: I1202 02:22:58.112198 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84531916-5cd5-425c-b024-a657aa80a781-catalog-content\") pod \"community-operators-8gk4c\" (UID: \"84531916-5cd5-425c-b024-a657aa80a781\") " pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:22:58 crc kubenswrapper[4884]: I1202 02:22:58.112487 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84531916-5cd5-425c-b024-a657aa80a781-catalog-content\") pod \"community-operators-8gk4c\" (UID: \"84531916-5cd5-425c-b024-a657aa80a781\") " pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:22:58 crc kubenswrapper[4884]: I1202 02:22:58.135768 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhwlf\" (UniqueName: \"kubernetes.io/projected/84531916-5cd5-425c-b024-a657aa80a781-kube-api-access-nhwlf\") pod \"community-operators-8gk4c\" (UID: \"84531916-5cd5-425c-b024-a657aa80a781\") " pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:22:58 crc kubenswrapper[4884]: I1202 02:22:58.171407 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:22:58 crc kubenswrapper[4884]: I1202 02:22:58.711362 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8gk4c"] Dec 02 02:22:58 crc kubenswrapper[4884]: W1202 02:22:58.723224 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84531916_5cd5_425c_b024_a657aa80a781.slice/crio-8af54b787b495f186b73be673d5b67bcd6d490718eaf57b656ad97a79b3ad32b WatchSource:0}: Error finding container 8af54b787b495f186b73be673d5b67bcd6d490718eaf57b656ad97a79b3ad32b: Status 404 returned error can't find the container with id 8af54b787b495f186b73be673d5b67bcd6d490718eaf57b656ad97a79b3ad32b Dec 02 02:22:59 crc kubenswrapper[4884]: I1202 02:22:59.232622 4884 generic.go:334] "Generic (PLEG): container finished" podID="84531916-5cd5-425c-b024-a657aa80a781" containerID="a0615d1f8a93a73052679211829f4247e6e838366d6acca6ba3f7480132c91ed" exitCode=0 Dec 02 02:22:59 crc kubenswrapper[4884]: I1202 02:22:59.232679 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8gk4c" event={"ID":"84531916-5cd5-425c-b024-a657aa80a781","Type":"ContainerDied","Data":"a0615d1f8a93a73052679211829f4247e6e838366d6acca6ba3f7480132c91ed"} Dec 02 02:22:59 crc kubenswrapper[4884]: I1202 02:22:59.232710 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8gk4c" event={"ID":"84531916-5cd5-425c-b024-a657aa80a781","Type":"ContainerStarted","Data":"8af54b787b495f186b73be673d5b67bcd6d490718eaf57b656ad97a79b3ad32b"} Dec 02 02:22:59 crc kubenswrapper[4884]: I1202 02:22:59.236472 4884 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 02:23:00 crc kubenswrapper[4884]: I1202 02:23:00.246890 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8gk4c" event={"ID":"84531916-5cd5-425c-b024-a657aa80a781","Type":"ContainerStarted","Data":"72350da8c2c40dafb1e8d4689474c030cdb3cf82ecb90e44d14e5acb863ffef9"} Dec 02 02:23:01 crc kubenswrapper[4884]: I1202 02:23:01.255488 4884 generic.go:334] "Generic (PLEG): container finished" podID="84531916-5cd5-425c-b024-a657aa80a781" containerID="72350da8c2c40dafb1e8d4689474c030cdb3cf82ecb90e44d14e5acb863ffef9" exitCode=0 Dec 02 02:23:01 crc kubenswrapper[4884]: I1202 02:23:01.255550 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8gk4c" event={"ID":"84531916-5cd5-425c-b024-a657aa80a781","Type":"ContainerDied","Data":"72350da8c2c40dafb1e8d4689474c030cdb3cf82ecb90e44d14e5acb863ffef9"} Dec 02 02:23:02 crc kubenswrapper[4884]: I1202 02:23:02.266388 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8gk4c" event={"ID":"84531916-5cd5-425c-b024-a657aa80a781","Type":"ContainerStarted","Data":"a678167474a02bcd7a0034bf4c4e490fb7105c39266db8a74d400503a32c585a"} Dec 02 02:23:02 crc kubenswrapper[4884]: I1202 02:23:02.294277 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8gk4c" podStartSLOduration=2.677067211 podStartE2EDuration="5.294260232s" podCreationTimestamp="2025-12-02 02:22:57 +0000 UTC" firstStartedPulling="2025-12-02 02:22:59.236048257 +0000 UTC m=+2795.911885171" lastFinishedPulling="2025-12-02 02:23:01.853241308 +0000 UTC m=+2798.529078192" observedRunningTime="2025-12-02 02:23:02.285233444 +0000 UTC m=+2798.961070338" watchObservedRunningTime="2025-12-02 02:23:02.294260232 +0000 UTC m=+2798.970097116" Dec 02 02:23:08 crc kubenswrapper[4884]: I1202 02:23:08.171726 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:23:08 crc kubenswrapper[4884]: I1202 02:23:08.173424 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:23:08 crc kubenswrapper[4884]: I1202 02:23:08.236870 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:23:08 crc kubenswrapper[4884]: I1202 02:23:08.390598 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:23:08 crc kubenswrapper[4884]: I1202 02:23:08.474514 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8gk4c"] Dec 02 02:23:10 crc kubenswrapper[4884]: I1202 02:23:10.363007 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8gk4c" podUID="84531916-5cd5-425c-b024-a657aa80a781" containerName="registry-server" containerID="cri-o://a678167474a02bcd7a0034bf4c4e490fb7105c39266db8a74d400503a32c585a" gracePeriod=2 Dec 02 02:23:10 crc kubenswrapper[4884]: I1202 02:23:10.929236 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:23:10 crc kubenswrapper[4884]: I1202 02:23:10.997027 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84531916-5cd5-425c-b024-a657aa80a781-utilities\") pod \"84531916-5cd5-425c-b024-a657aa80a781\" (UID: \"84531916-5cd5-425c-b024-a657aa80a781\") " Dec 02 02:23:10 crc kubenswrapper[4884]: I1202 02:23:10.997143 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhwlf\" (UniqueName: \"kubernetes.io/projected/84531916-5cd5-425c-b024-a657aa80a781-kube-api-access-nhwlf\") pod \"84531916-5cd5-425c-b024-a657aa80a781\" (UID: \"84531916-5cd5-425c-b024-a657aa80a781\") " Dec 02 02:23:10 crc kubenswrapper[4884]: I1202 02:23:10.997201 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84531916-5cd5-425c-b024-a657aa80a781-catalog-content\") pod \"84531916-5cd5-425c-b024-a657aa80a781\" (UID: \"84531916-5cd5-425c-b024-a657aa80a781\") " Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:10.999554 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84531916-5cd5-425c-b024-a657aa80a781-utilities" (OuterVolumeSpecName: "utilities") pod "84531916-5cd5-425c-b024-a657aa80a781" (UID: "84531916-5cd5-425c-b024-a657aa80a781"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.005073 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84531916-5cd5-425c-b024-a657aa80a781-kube-api-access-nhwlf" (OuterVolumeSpecName: "kube-api-access-nhwlf") pod "84531916-5cd5-425c-b024-a657aa80a781" (UID: "84531916-5cd5-425c-b024-a657aa80a781"). InnerVolumeSpecName "kube-api-access-nhwlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.099896 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84531916-5cd5-425c-b024-a657aa80a781-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.099923 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhwlf\" (UniqueName: \"kubernetes.io/projected/84531916-5cd5-425c-b024-a657aa80a781-kube-api-access-nhwlf\") on node \"crc\" DevicePath \"\"" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.174593 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84531916-5cd5-425c-b024-a657aa80a781-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84531916-5cd5-425c-b024-a657aa80a781" (UID: "84531916-5cd5-425c-b024-a657aa80a781"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.202032 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84531916-5cd5-425c-b024-a657aa80a781-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.379411 4884 generic.go:334] "Generic (PLEG): container finished" podID="84531916-5cd5-425c-b024-a657aa80a781" containerID="a678167474a02bcd7a0034bf4c4e490fb7105c39266db8a74d400503a32c585a" exitCode=0 Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.379458 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8gk4c" event={"ID":"84531916-5cd5-425c-b024-a657aa80a781","Type":"ContainerDied","Data":"a678167474a02bcd7a0034bf4c4e490fb7105c39266db8a74d400503a32c585a"} Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.379498 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8gk4c" event={"ID":"84531916-5cd5-425c-b024-a657aa80a781","Type":"ContainerDied","Data":"8af54b787b495f186b73be673d5b67bcd6d490718eaf57b656ad97a79b3ad32b"} Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.379506 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8gk4c" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.379526 4884 scope.go:117] "RemoveContainer" containerID="a678167474a02bcd7a0034bf4c4e490fb7105c39266db8a74d400503a32c585a" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.421533 4884 scope.go:117] "RemoveContainer" containerID="72350da8c2c40dafb1e8d4689474c030cdb3cf82ecb90e44d14e5acb863ffef9" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.444951 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8gk4c"] Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.457283 4884 scope.go:117] "RemoveContainer" containerID="a0615d1f8a93a73052679211829f4247e6e838366d6acca6ba3f7480132c91ed" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.458856 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8gk4c"] Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.546675 4884 scope.go:117] "RemoveContainer" containerID="a678167474a02bcd7a0034bf4c4e490fb7105c39266db8a74d400503a32c585a" Dec 02 02:23:11 crc kubenswrapper[4884]: E1202 02:23:11.551237 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a678167474a02bcd7a0034bf4c4e490fb7105c39266db8a74d400503a32c585a\": container with ID starting with a678167474a02bcd7a0034bf4c4e490fb7105c39266db8a74d400503a32c585a not found: ID does not exist" containerID="a678167474a02bcd7a0034bf4c4e490fb7105c39266db8a74d400503a32c585a" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.551286 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a678167474a02bcd7a0034bf4c4e490fb7105c39266db8a74d400503a32c585a"} err="failed to get container status \"a678167474a02bcd7a0034bf4c4e490fb7105c39266db8a74d400503a32c585a\": rpc error: code = NotFound desc = could not find container \"a678167474a02bcd7a0034bf4c4e490fb7105c39266db8a74d400503a32c585a\": container with ID starting with a678167474a02bcd7a0034bf4c4e490fb7105c39266db8a74d400503a32c585a not found: ID does not exist" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.551318 4884 scope.go:117] "RemoveContainer" containerID="72350da8c2c40dafb1e8d4689474c030cdb3cf82ecb90e44d14e5acb863ffef9" Dec 02 02:23:11 crc kubenswrapper[4884]: E1202 02:23:11.551819 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72350da8c2c40dafb1e8d4689474c030cdb3cf82ecb90e44d14e5acb863ffef9\": container with ID starting with 72350da8c2c40dafb1e8d4689474c030cdb3cf82ecb90e44d14e5acb863ffef9 not found: ID does not exist" containerID="72350da8c2c40dafb1e8d4689474c030cdb3cf82ecb90e44d14e5acb863ffef9" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.551849 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72350da8c2c40dafb1e8d4689474c030cdb3cf82ecb90e44d14e5acb863ffef9"} err="failed to get container status \"72350da8c2c40dafb1e8d4689474c030cdb3cf82ecb90e44d14e5acb863ffef9\": rpc error: code = NotFound desc = could not find container \"72350da8c2c40dafb1e8d4689474c030cdb3cf82ecb90e44d14e5acb863ffef9\": container with ID starting with 72350da8c2c40dafb1e8d4689474c030cdb3cf82ecb90e44d14e5acb863ffef9 not found: ID does not exist" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.551867 4884 scope.go:117] "RemoveContainer" containerID="a0615d1f8a93a73052679211829f4247e6e838366d6acca6ba3f7480132c91ed" Dec 02 02:23:11 crc kubenswrapper[4884]: E1202 02:23:11.552438 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0615d1f8a93a73052679211829f4247e6e838366d6acca6ba3f7480132c91ed\": container with ID starting with a0615d1f8a93a73052679211829f4247e6e838366d6acca6ba3f7480132c91ed not found: ID does not exist" containerID="a0615d1f8a93a73052679211829f4247e6e838366d6acca6ba3f7480132c91ed" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.552496 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0615d1f8a93a73052679211829f4247e6e838366d6acca6ba3f7480132c91ed"} err="failed to get container status \"a0615d1f8a93a73052679211829f4247e6e838366d6acca6ba3f7480132c91ed\": rpc error: code = NotFound desc = could not find container \"a0615d1f8a93a73052679211829f4247e6e838366d6acca6ba3f7480132c91ed\": container with ID starting with a0615d1f8a93a73052679211829f4247e6e838366d6acca6ba3f7480132c91ed not found: ID does not exist" Dec 02 02:23:11 crc kubenswrapper[4884]: I1202 02:23:11.634118 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84531916-5cd5-425c-b024-a657aa80a781" path="/var/lib/kubelet/pods/84531916-5cd5-425c-b024-a657aa80a781/volumes" Dec 02 02:23:46 crc kubenswrapper[4884]: I1202 02:23:46.971723 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:23:46 crc kubenswrapper[4884]: I1202 02:23:46.972266 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:23:54 crc kubenswrapper[4884]: I1202 02:23:54.870376 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gpk62"] Dec 02 02:23:54 crc kubenswrapper[4884]: E1202 02:23:54.871488 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84531916-5cd5-425c-b024-a657aa80a781" containerName="extract-content" Dec 02 02:23:54 crc kubenswrapper[4884]: I1202 02:23:54.871508 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="84531916-5cd5-425c-b024-a657aa80a781" containerName="extract-content" Dec 02 02:23:54 crc kubenswrapper[4884]: E1202 02:23:54.871544 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84531916-5cd5-425c-b024-a657aa80a781" containerName="registry-server" Dec 02 02:23:54 crc kubenswrapper[4884]: I1202 02:23:54.871553 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="84531916-5cd5-425c-b024-a657aa80a781" containerName="registry-server" Dec 02 02:23:54 crc kubenswrapper[4884]: E1202 02:23:54.871568 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84531916-5cd5-425c-b024-a657aa80a781" containerName="extract-utilities" Dec 02 02:23:54 crc kubenswrapper[4884]: I1202 02:23:54.871576 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="84531916-5cd5-425c-b024-a657aa80a781" containerName="extract-utilities" Dec 02 02:23:54 crc kubenswrapper[4884]: I1202 02:23:54.871852 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="84531916-5cd5-425c-b024-a657aa80a781" containerName="registry-server" Dec 02 02:23:54 crc kubenswrapper[4884]: I1202 02:23:54.873716 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:23:54 crc kubenswrapper[4884]: I1202 02:23:54.883772 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gpk62"] Dec 02 02:23:54 crc kubenswrapper[4884]: I1202 02:23:54.999514 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4xrr\" (UniqueName: \"kubernetes.io/projected/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-kube-api-access-t4xrr\") pod \"redhat-marketplace-gpk62\" (UID: \"a0f8c73f-3d21-4acf-9c1e-5a73908ed387\") " pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:23:54 crc kubenswrapper[4884]: I1202 02:23:54.999566 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-catalog-content\") pod \"redhat-marketplace-gpk62\" (UID: \"a0f8c73f-3d21-4acf-9c1e-5a73908ed387\") " pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:23:55 crc kubenswrapper[4884]: I1202 02:23:54.999970 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-utilities\") pod \"redhat-marketplace-gpk62\" (UID: \"a0f8c73f-3d21-4acf-9c1e-5a73908ed387\") " pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:23:55 crc kubenswrapper[4884]: I1202 02:23:55.101911 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-utilities\") pod \"redhat-marketplace-gpk62\" (UID: \"a0f8c73f-3d21-4acf-9c1e-5a73908ed387\") " pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:23:55 crc kubenswrapper[4884]: I1202 02:23:55.102005 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4xrr\" (UniqueName: \"kubernetes.io/projected/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-kube-api-access-t4xrr\") pod \"redhat-marketplace-gpk62\" (UID: \"a0f8c73f-3d21-4acf-9c1e-5a73908ed387\") " pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:23:55 crc kubenswrapper[4884]: I1202 02:23:55.102044 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-catalog-content\") pod \"redhat-marketplace-gpk62\" (UID: \"a0f8c73f-3d21-4acf-9c1e-5a73908ed387\") " pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:23:55 crc kubenswrapper[4884]: I1202 02:23:55.102436 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-utilities\") pod \"redhat-marketplace-gpk62\" (UID: \"a0f8c73f-3d21-4acf-9c1e-5a73908ed387\") " pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:23:55 crc kubenswrapper[4884]: I1202 02:23:55.102501 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-catalog-content\") pod \"redhat-marketplace-gpk62\" (UID: \"a0f8c73f-3d21-4acf-9c1e-5a73908ed387\") " pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:23:55 crc kubenswrapper[4884]: I1202 02:23:55.123083 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4xrr\" (UniqueName: \"kubernetes.io/projected/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-kube-api-access-t4xrr\") pod \"redhat-marketplace-gpk62\" (UID: \"a0f8c73f-3d21-4acf-9c1e-5a73908ed387\") " pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:23:55 crc kubenswrapper[4884]: I1202 02:23:55.216813 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:23:55 crc kubenswrapper[4884]: I1202 02:23:55.793046 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gpk62"] Dec 02 02:23:55 crc kubenswrapper[4884]: I1202 02:23:55.954111 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpk62" event={"ID":"a0f8c73f-3d21-4acf-9c1e-5a73908ed387","Type":"ContainerStarted","Data":"76bda38c08e09d9bdca24120245c6cb101e7bc1874fda6119bf20af52df4eea0"} Dec 02 02:23:56 crc kubenswrapper[4884]: I1202 02:23:56.968960 4884 generic.go:334] "Generic (PLEG): container finished" podID="a0f8c73f-3d21-4acf-9c1e-5a73908ed387" containerID="f3f9558d2a2cfb5623274440e2bc1bb7aa5d9c6eeca548bdd03e4444defd0cac" exitCode=0 Dec 02 02:23:56 crc kubenswrapper[4884]: I1202 02:23:56.969063 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpk62" event={"ID":"a0f8c73f-3d21-4acf-9c1e-5a73908ed387","Type":"ContainerDied","Data":"f3f9558d2a2cfb5623274440e2bc1bb7aa5d9c6eeca548bdd03e4444defd0cac"} Dec 02 02:23:59 crc kubenswrapper[4884]: I1202 02:23:59.007852 4884 generic.go:334] "Generic (PLEG): container finished" podID="a0f8c73f-3d21-4acf-9c1e-5a73908ed387" containerID="d83e2deeb40162c8b9018c730cb38d4994b6330a59d6513a304a3a77e32d9724" exitCode=0 Dec 02 02:23:59 crc kubenswrapper[4884]: I1202 02:23:59.007909 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpk62" event={"ID":"a0f8c73f-3d21-4acf-9c1e-5a73908ed387","Type":"ContainerDied","Data":"d83e2deeb40162c8b9018c730cb38d4994b6330a59d6513a304a3a77e32d9724"} Dec 02 02:24:00 crc kubenswrapper[4884]: I1202 02:24:00.020190 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpk62" event={"ID":"a0f8c73f-3d21-4acf-9c1e-5a73908ed387","Type":"ContainerStarted","Data":"73d0a29b373988aa9fb3bfa5ce22949602eeb90da9409708010943e6b29bf790"} Dec 02 02:24:00 crc kubenswrapper[4884]: I1202 02:24:00.048635 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gpk62" podStartSLOduration=3.452222269 podStartE2EDuration="6.048617129s" podCreationTimestamp="2025-12-02 02:23:54 +0000 UTC" firstStartedPulling="2025-12-02 02:23:56.971436564 +0000 UTC m=+2853.647273488" lastFinishedPulling="2025-12-02 02:23:59.567831424 +0000 UTC m=+2856.243668348" observedRunningTime="2025-12-02 02:24:00.044048308 +0000 UTC m=+2856.719885212" watchObservedRunningTime="2025-12-02 02:24:00.048617129 +0000 UTC m=+2856.724454023" Dec 02 02:24:05 crc kubenswrapper[4884]: I1202 02:24:05.217574 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:24:05 crc kubenswrapper[4884]: I1202 02:24:05.218425 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:24:05 crc kubenswrapper[4884]: I1202 02:24:05.294297 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:24:06 crc kubenswrapper[4884]: I1202 02:24:06.156968 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:24:06 crc kubenswrapper[4884]: I1202 02:24:06.235649 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gpk62"] Dec 02 02:24:08 crc kubenswrapper[4884]: I1202 02:24:08.112521 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gpk62" podUID="a0f8c73f-3d21-4acf-9c1e-5a73908ed387" containerName="registry-server" containerID="cri-o://73d0a29b373988aa9fb3bfa5ce22949602eeb90da9409708010943e6b29bf790" gracePeriod=2 Dec 02 02:24:08 crc kubenswrapper[4884]: I1202 02:24:08.679067 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:24:08 crc kubenswrapper[4884]: I1202 02:24:08.823349 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4xrr\" (UniqueName: \"kubernetes.io/projected/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-kube-api-access-t4xrr\") pod \"a0f8c73f-3d21-4acf-9c1e-5a73908ed387\" (UID: \"a0f8c73f-3d21-4acf-9c1e-5a73908ed387\") " Dec 02 02:24:08 crc kubenswrapper[4884]: I1202 02:24:08.823488 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-catalog-content\") pod \"a0f8c73f-3d21-4acf-9c1e-5a73908ed387\" (UID: \"a0f8c73f-3d21-4acf-9c1e-5a73908ed387\") " Dec 02 02:24:08 crc kubenswrapper[4884]: I1202 02:24:08.823533 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-utilities\") pod \"a0f8c73f-3d21-4acf-9c1e-5a73908ed387\" (UID: \"a0f8c73f-3d21-4acf-9c1e-5a73908ed387\") " Dec 02 02:24:08 crc kubenswrapper[4884]: I1202 02:24:08.824566 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-utilities" (OuterVolumeSpecName: "utilities") pod "a0f8c73f-3d21-4acf-9c1e-5a73908ed387" (UID: "a0f8c73f-3d21-4acf-9c1e-5a73908ed387"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:24:08 crc kubenswrapper[4884]: I1202 02:24:08.830058 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-kube-api-access-t4xrr" (OuterVolumeSpecName: "kube-api-access-t4xrr") pod "a0f8c73f-3d21-4acf-9c1e-5a73908ed387" (UID: "a0f8c73f-3d21-4acf-9c1e-5a73908ed387"). InnerVolumeSpecName "kube-api-access-t4xrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:24:08 crc kubenswrapper[4884]: I1202 02:24:08.843786 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0f8c73f-3d21-4acf-9c1e-5a73908ed387" (UID: "a0f8c73f-3d21-4acf-9c1e-5a73908ed387"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:24:08 crc kubenswrapper[4884]: I1202 02:24:08.926080 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:24:08 crc kubenswrapper[4884]: I1202 02:24:08.926113 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:24:08 crc kubenswrapper[4884]: I1202 02:24:08.926124 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4xrr\" (UniqueName: \"kubernetes.io/projected/a0f8c73f-3d21-4acf-9c1e-5a73908ed387-kube-api-access-t4xrr\") on node \"crc\" DevicePath \"\"" Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.128519 4884 generic.go:334] "Generic (PLEG): container finished" podID="a0f8c73f-3d21-4acf-9c1e-5a73908ed387" containerID="73d0a29b373988aa9fb3bfa5ce22949602eeb90da9409708010943e6b29bf790" exitCode=0 Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.128635 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpk62" event={"ID":"a0f8c73f-3d21-4acf-9c1e-5a73908ed387","Type":"ContainerDied","Data":"73d0a29b373988aa9fb3bfa5ce22949602eeb90da9409708010943e6b29bf790"} Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.128965 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gpk62" event={"ID":"a0f8c73f-3d21-4acf-9c1e-5a73908ed387","Type":"ContainerDied","Data":"76bda38c08e09d9bdca24120245c6cb101e7bc1874fda6119bf20af52df4eea0"} Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.128650 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gpk62" Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.129003 4884 scope.go:117] "RemoveContainer" containerID="73d0a29b373988aa9fb3bfa5ce22949602eeb90da9409708010943e6b29bf790" Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.169464 4884 scope.go:117] "RemoveContainer" containerID="d83e2deeb40162c8b9018c730cb38d4994b6330a59d6513a304a3a77e32d9724" Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.173675 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gpk62"] Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.183792 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gpk62"] Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.202024 4884 scope.go:117] "RemoveContainer" containerID="f3f9558d2a2cfb5623274440e2bc1bb7aa5d9c6eeca548bdd03e4444defd0cac" Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.259960 4884 scope.go:117] "RemoveContainer" containerID="73d0a29b373988aa9fb3bfa5ce22949602eeb90da9409708010943e6b29bf790" Dec 02 02:24:09 crc kubenswrapper[4884]: E1202 02:24:09.260659 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73d0a29b373988aa9fb3bfa5ce22949602eeb90da9409708010943e6b29bf790\": container with ID starting with 73d0a29b373988aa9fb3bfa5ce22949602eeb90da9409708010943e6b29bf790 not found: ID does not exist" containerID="73d0a29b373988aa9fb3bfa5ce22949602eeb90da9409708010943e6b29bf790" Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.260708 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73d0a29b373988aa9fb3bfa5ce22949602eeb90da9409708010943e6b29bf790"} err="failed to get container status \"73d0a29b373988aa9fb3bfa5ce22949602eeb90da9409708010943e6b29bf790\": rpc error: code = NotFound desc = could not find container \"73d0a29b373988aa9fb3bfa5ce22949602eeb90da9409708010943e6b29bf790\": container with ID starting with 73d0a29b373988aa9fb3bfa5ce22949602eeb90da9409708010943e6b29bf790 not found: ID does not exist" Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.260759 4884 scope.go:117] "RemoveContainer" containerID="d83e2deeb40162c8b9018c730cb38d4994b6330a59d6513a304a3a77e32d9724" Dec 02 02:24:09 crc kubenswrapper[4884]: E1202 02:24:09.261217 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d83e2deeb40162c8b9018c730cb38d4994b6330a59d6513a304a3a77e32d9724\": container with ID starting with d83e2deeb40162c8b9018c730cb38d4994b6330a59d6513a304a3a77e32d9724 not found: ID does not exist" containerID="d83e2deeb40162c8b9018c730cb38d4994b6330a59d6513a304a3a77e32d9724" Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.261257 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d83e2deeb40162c8b9018c730cb38d4994b6330a59d6513a304a3a77e32d9724"} err="failed to get container status \"d83e2deeb40162c8b9018c730cb38d4994b6330a59d6513a304a3a77e32d9724\": rpc error: code = NotFound desc = could not find container \"d83e2deeb40162c8b9018c730cb38d4994b6330a59d6513a304a3a77e32d9724\": container with ID starting with d83e2deeb40162c8b9018c730cb38d4994b6330a59d6513a304a3a77e32d9724 not found: ID does not exist" Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.261285 4884 scope.go:117] "RemoveContainer" containerID="f3f9558d2a2cfb5623274440e2bc1bb7aa5d9c6eeca548bdd03e4444defd0cac" Dec 02 02:24:09 crc kubenswrapper[4884]: E1202 02:24:09.262430 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3f9558d2a2cfb5623274440e2bc1bb7aa5d9c6eeca548bdd03e4444defd0cac\": container with ID starting with f3f9558d2a2cfb5623274440e2bc1bb7aa5d9c6eeca548bdd03e4444defd0cac not found: ID does not exist" containerID="f3f9558d2a2cfb5623274440e2bc1bb7aa5d9c6eeca548bdd03e4444defd0cac" Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.262461 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3f9558d2a2cfb5623274440e2bc1bb7aa5d9c6eeca548bdd03e4444defd0cac"} err="failed to get container status \"f3f9558d2a2cfb5623274440e2bc1bb7aa5d9c6eeca548bdd03e4444defd0cac\": rpc error: code = NotFound desc = could not find container \"f3f9558d2a2cfb5623274440e2bc1bb7aa5d9c6eeca548bdd03e4444defd0cac\": container with ID starting with f3f9558d2a2cfb5623274440e2bc1bb7aa5d9c6eeca548bdd03e4444defd0cac not found: ID does not exist" Dec 02 02:24:09 crc kubenswrapper[4884]: I1202 02:24:09.638998 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0f8c73f-3d21-4acf-9c1e-5a73908ed387" path="/var/lib/kubelet/pods/a0f8c73f-3d21-4acf-9c1e-5a73908ed387/volumes" Dec 02 02:24:11 crc kubenswrapper[4884]: I1202 02:24:11.177937 4884 generic.go:334] "Generic (PLEG): container finished" podID="ba6d4723-02c0-46ce-bfd0-afa42fd31d7d" containerID="576bb405ff77b646151ed41939d6aff34d31421f203dd64f897404f70f5437c6" exitCode=0 Dec 02 02:24:11 crc kubenswrapper[4884]: I1202 02:24:11.178075 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" event={"ID":"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d","Type":"ContainerDied","Data":"576bb405ff77b646151ed41939d6aff34d31421f203dd64f897404f70f5437c6"} Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.727203 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.829505 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-ssh-key\") pod \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.829897 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-libvirt-secret-0\") pod \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.830019 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6pr8\" (UniqueName: \"kubernetes.io/projected/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-kube-api-access-p6pr8\") pod \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.830072 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-libvirt-combined-ca-bundle\") pod \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.830120 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-inventory\") pod \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\" (UID: \"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d\") " Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.836697 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "ba6d4723-02c0-46ce-bfd0-afa42fd31d7d" (UID: "ba6d4723-02c0-46ce-bfd0-afa42fd31d7d"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.839411 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-kube-api-access-p6pr8" (OuterVolumeSpecName: "kube-api-access-p6pr8") pod "ba6d4723-02c0-46ce-bfd0-afa42fd31d7d" (UID: "ba6d4723-02c0-46ce-bfd0-afa42fd31d7d"). InnerVolumeSpecName "kube-api-access-p6pr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.874194 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-inventory" (OuterVolumeSpecName: "inventory") pod "ba6d4723-02c0-46ce-bfd0-afa42fd31d7d" (UID: "ba6d4723-02c0-46ce-bfd0-afa42fd31d7d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.880319 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "ba6d4723-02c0-46ce-bfd0-afa42fd31d7d" (UID: "ba6d4723-02c0-46ce-bfd0-afa42fd31d7d"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.888506 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ba6d4723-02c0-46ce-bfd0-afa42fd31d7d" (UID: "ba6d4723-02c0-46ce-bfd0-afa42fd31d7d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.932411 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.932473 4884 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.932496 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6pr8\" (UniqueName: \"kubernetes.io/projected/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-kube-api-access-p6pr8\") on node \"crc\" DevicePath \"\"" Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.932515 4884 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:24:12 crc kubenswrapper[4884]: I1202 02:24:12.932534 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ba6d4723-02c0-46ce-bfd0-afa42fd31d7d-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.201374 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" event={"ID":"ba6d4723-02c0-46ce-bfd0-afa42fd31d7d","Type":"ContainerDied","Data":"af2aeeb20547604a60afc3e34fde38f15596b0742ae588cd8149953a4a954fdc"} Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.201427 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af2aeeb20547604a60afc3e34fde38f15596b0742ae588cd8149953a4a954fdc" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.201447 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.342646 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb"] Dec 02 02:24:13 crc kubenswrapper[4884]: E1202 02:24:13.343214 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba6d4723-02c0-46ce-bfd0-afa42fd31d7d" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.343236 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6d4723-02c0-46ce-bfd0-afa42fd31d7d" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 02 02:24:13 crc kubenswrapper[4884]: E1202 02:24:13.343253 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f8c73f-3d21-4acf-9c1e-5a73908ed387" containerName="registry-server" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.343262 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f8c73f-3d21-4acf-9c1e-5a73908ed387" containerName="registry-server" Dec 02 02:24:13 crc kubenswrapper[4884]: E1202 02:24:13.343284 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f8c73f-3d21-4acf-9c1e-5a73908ed387" containerName="extract-content" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.343292 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f8c73f-3d21-4acf-9c1e-5a73908ed387" containerName="extract-content" Dec 02 02:24:13 crc kubenswrapper[4884]: E1202 02:24:13.343319 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f8c73f-3d21-4acf-9c1e-5a73908ed387" containerName="extract-utilities" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.343328 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f8c73f-3d21-4acf-9c1e-5a73908ed387" containerName="extract-utilities" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.343567 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0f8c73f-3d21-4acf-9c1e-5a73908ed387" containerName="registry-server" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.343587 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba6d4723-02c0-46ce-bfd0-afa42fd31d7d" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.344528 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.346331 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.347593 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.347708 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.347880 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.347598 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.354331 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.354260 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.361999 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb"] Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.458768 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.459100 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkpx4\" (UniqueName: \"kubernetes.io/projected/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-kube-api-access-dkpx4\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.459131 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.459189 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.459244 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.459296 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.459345 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.459538 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.459602 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.561937 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.562003 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.562053 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.562107 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.562136 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.562167 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.562261 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.562290 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkpx4\" (UniqueName: \"kubernetes.io/projected/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-kube-api-access-dkpx4\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.562314 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.563134 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.565783 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.566402 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.568469 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.569408 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.569952 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.570222 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.577401 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.578027 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkpx4\" (UniqueName: \"kubernetes.io/projected/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-kube-api-access-dkpx4\") pod \"nova-edpm-deployment-openstack-edpm-ipam-lqsdb\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:13 crc kubenswrapper[4884]: I1202 02:24:13.666073 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:24:14 crc kubenswrapper[4884]: W1202 02:24:14.249722 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ee4fe1d_9395_4e57_8d2c_7dca61af8238.slice/crio-1cd5a42adc614b8887ae99a94b8df3428dc59bdc4a2dd37695c12ebdd4cbb0c0 WatchSource:0}: Error finding container 1cd5a42adc614b8887ae99a94b8df3428dc59bdc4a2dd37695c12ebdd4cbb0c0: Status 404 returned error can't find the container with id 1cd5a42adc614b8887ae99a94b8df3428dc59bdc4a2dd37695c12ebdd4cbb0c0 Dec 02 02:24:14 crc kubenswrapper[4884]: I1202 02:24:14.259123 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb"] Dec 02 02:24:15 crc kubenswrapper[4884]: I1202 02:24:15.226431 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" event={"ID":"7ee4fe1d-9395-4e57-8d2c-7dca61af8238","Type":"ContainerStarted","Data":"7c581371840b2a5faaa9adfe7613acff4c31080c75a9c4d375ec7ccd05d0ddde"} Dec 02 02:24:15 crc kubenswrapper[4884]: I1202 02:24:15.227327 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" event={"ID":"7ee4fe1d-9395-4e57-8d2c-7dca61af8238","Type":"ContainerStarted","Data":"1cd5a42adc614b8887ae99a94b8df3428dc59bdc4a2dd37695c12ebdd4cbb0c0"} Dec 02 02:24:15 crc kubenswrapper[4884]: I1202 02:24:15.261772 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" podStartSLOduration=1.735112617 podStartE2EDuration="2.261739419s" podCreationTimestamp="2025-12-02 02:24:13 +0000 UTC" firstStartedPulling="2025-12-02 02:24:14.252292362 +0000 UTC m=+2870.928129246" lastFinishedPulling="2025-12-02 02:24:14.778919134 +0000 UTC m=+2871.454756048" observedRunningTime="2025-12-02 02:24:15.261587555 +0000 UTC m=+2871.937424479" watchObservedRunningTime="2025-12-02 02:24:15.261739419 +0000 UTC m=+2871.937576303" Dec 02 02:24:16 crc kubenswrapper[4884]: I1202 02:24:16.972243 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:24:16 crc kubenswrapper[4884]: I1202 02:24:16.972724 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:24:46 crc kubenswrapper[4884]: I1202 02:24:46.972378 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:24:46 crc kubenswrapper[4884]: I1202 02:24:46.973233 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:24:46 crc kubenswrapper[4884]: I1202 02:24:46.973306 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 02:24:46 crc kubenswrapper[4884]: I1202 02:24:46.974489 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"83598bb7142be85f0e3373487befb0fe500a455b9b41c9b4af785be9d73c5ae8"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 02:24:46 crc kubenswrapper[4884]: I1202 02:24:46.974611 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://83598bb7142be85f0e3373487befb0fe500a455b9b41c9b4af785be9d73c5ae8" gracePeriod=600 Dec 02 02:24:47 crc kubenswrapper[4884]: I1202 02:24:47.631008 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="83598bb7142be85f0e3373487befb0fe500a455b9b41c9b4af785be9d73c5ae8" exitCode=0 Dec 02 02:24:47 crc kubenswrapper[4884]: I1202 02:24:47.631648 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"83598bb7142be85f0e3373487befb0fe500a455b9b41c9b4af785be9d73c5ae8"} Dec 02 02:24:47 crc kubenswrapper[4884]: I1202 02:24:47.631740 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9"} Dec 02 02:24:47 crc kubenswrapper[4884]: I1202 02:24:47.631801 4884 scope.go:117] "RemoveContainer" containerID="a86ceea330675ec4502d43ead693e474b197c3e7f0c0ebfb66e6ff999dece653" Dec 02 02:27:16 crc kubenswrapper[4884]: I1202 02:27:16.972211 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:27:16 crc kubenswrapper[4884]: I1202 02:27:16.973077 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:27:42 crc kubenswrapper[4884]: I1202 02:27:42.761589 4884 generic.go:334] "Generic (PLEG): container finished" podID="7ee4fe1d-9395-4e57-8d2c-7dca61af8238" containerID="7c581371840b2a5faaa9adfe7613acff4c31080c75a9c4d375ec7ccd05d0ddde" exitCode=0 Dec 02 02:27:42 crc kubenswrapper[4884]: I1202 02:27:42.761660 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" event={"ID":"7ee4fe1d-9395-4e57-8d2c-7dca61af8238","Type":"ContainerDied","Data":"7c581371840b2a5faaa9adfe7613acff4c31080c75a9c4d375ec7ccd05d0ddde"} Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.256248 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.336677 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-cell1-compute-config-1\") pod \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.336814 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-inventory\") pod \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.336920 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-extra-config-0\") pod \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.336957 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkpx4\" (UniqueName: \"kubernetes.io/projected/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-kube-api-access-dkpx4\") pod \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.336998 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-migration-ssh-key-0\") pod \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.337071 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-cell1-compute-config-0\") pod \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.337151 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-combined-ca-bundle\") pod \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.337195 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-ssh-key\") pod \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.337225 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-migration-ssh-key-1\") pod \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\" (UID: \"7ee4fe1d-9395-4e57-8d2c-7dca61af8238\") " Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.345357 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-kube-api-access-dkpx4" (OuterVolumeSpecName: "kube-api-access-dkpx4") pod "7ee4fe1d-9395-4e57-8d2c-7dca61af8238" (UID: "7ee4fe1d-9395-4e57-8d2c-7dca61af8238"). InnerVolumeSpecName "kube-api-access-dkpx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.348493 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "7ee4fe1d-9395-4e57-8d2c-7dca61af8238" (UID: "7ee4fe1d-9395-4e57-8d2c-7dca61af8238"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.370708 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "7ee4fe1d-9395-4e57-8d2c-7dca61af8238" (UID: "7ee4fe1d-9395-4e57-8d2c-7dca61af8238"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.371199 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-inventory" (OuterVolumeSpecName: "inventory") pod "7ee4fe1d-9395-4e57-8d2c-7dca61af8238" (UID: "7ee4fe1d-9395-4e57-8d2c-7dca61af8238"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.373213 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "7ee4fe1d-9395-4e57-8d2c-7dca61af8238" (UID: "7ee4fe1d-9395-4e57-8d2c-7dca61af8238"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.375876 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7ee4fe1d-9395-4e57-8d2c-7dca61af8238" (UID: "7ee4fe1d-9395-4e57-8d2c-7dca61af8238"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.376479 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "7ee4fe1d-9395-4e57-8d2c-7dca61af8238" (UID: "7ee4fe1d-9395-4e57-8d2c-7dca61af8238"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.387534 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "7ee4fe1d-9395-4e57-8d2c-7dca61af8238" (UID: "7ee4fe1d-9395-4e57-8d2c-7dca61af8238"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.395086 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "7ee4fe1d-9395-4e57-8d2c-7dca61af8238" (UID: "7ee4fe1d-9395-4e57-8d2c-7dca61af8238"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.440009 4884 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.440048 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.440061 4884 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.440074 4884 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.440089 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.440105 4884 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.440118 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkpx4\" (UniqueName: \"kubernetes.io/projected/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-kube-api-access-dkpx4\") on node \"crc\" DevicePath \"\"" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.440130 4884 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.440141 4884 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7ee4fe1d-9395-4e57-8d2c-7dca61af8238-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.787565 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" event={"ID":"7ee4fe1d-9395-4e57-8d2c-7dca61af8238","Type":"ContainerDied","Data":"1cd5a42adc614b8887ae99a94b8df3428dc59bdc4a2dd37695c12ebdd4cbb0c0"} Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.787604 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cd5a42adc614b8887ae99a94b8df3428dc59bdc4a2dd37695c12ebdd4cbb0c0" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.787694 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-lqsdb" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.938294 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz"] Dec 02 02:27:44 crc kubenswrapper[4884]: E1202 02:27:44.939622 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee4fe1d-9395-4e57-8d2c-7dca61af8238" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.942269 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee4fe1d-9395-4e57-8d2c-7dca61af8238" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.943093 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ee4fe1d-9395-4e57-8d2c-7dca61af8238" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.944977 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.948111 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.948429 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.948625 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.948850 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.950086 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-ckhf2" Dec 02 02:27:44 crc kubenswrapper[4884]: I1202 02:27:44.955298 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz"] Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.052682 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.052780 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.052874 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.052935 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.052996 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.053022 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.053058 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx62g\" (UniqueName: \"kubernetes.io/projected/7098a534-9f0a-457d-a3d2-09249bd490ee-kube-api-access-jx62g\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.154991 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.155407 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.155447 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.155494 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx62g\" (UniqueName: \"kubernetes.io/projected/7098a534-9f0a-457d-a3d2-09249bd490ee-kube-api-access-jx62g\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.155623 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.155666 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.155760 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.159549 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.159999 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.160049 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.167556 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.170789 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.178572 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.184885 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx62g\" (UniqueName: \"kubernetes.io/projected/7098a534-9f0a-457d-a3d2-09249bd490ee-kube-api-access-jx62g\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.276007 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:27:45 crc kubenswrapper[4884]: I1202 02:27:45.869332 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz"] Dec 02 02:27:46 crc kubenswrapper[4884]: I1202 02:27:46.815543 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" event={"ID":"7098a534-9f0a-457d-a3d2-09249bd490ee","Type":"ContainerStarted","Data":"bf2bc2c42e09fd749876f3c7a24faad004820eebf2a3518f90aa1b5017b9a131"} Dec 02 02:27:46 crc kubenswrapper[4884]: I1202 02:27:46.816384 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" event={"ID":"7098a534-9f0a-457d-a3d2-09249bd490ee","Type":"ContainerStarted","Data":"4b28e939d456cd51cabb331a6c32828b973e9c1e626af1b11e9295c78bbe13b5"} Dec 02 02:27:46 crc kubenswrapper[4884]: I1202 02:27:46.842433 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" podStartSLOduration=2.383313723 podStartE2EDuration="2.842405064s" podCreationTimestamp="2025-12-02 02:27:44 +0000 UTC" firstStartedPulling="2025-12-02 02:27:45.878769884 +0000 UTC m=+3082.554606808" lastFinishedPulling="2025-12-02 02:27:46.337861255 +0000 UTC m=+3083.013698149" observedRunningTime="2025-12-02 02:27:46.841641706 +0000 UTC m=+3083.517478660" watchObservedRunningTime="2025-12-02 02:27:46.842405064 +0000 UTC m=+3083.518241978" Dec 02 02:27:46 crc kubenswrapper[4884]: I1202 02:27:46.972260 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:27:46 crc kubenswrapper[4884]: I1202 02:27:46.972351 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:28:00 crc kubenswrapper[4884]: I1202 02:28:00.703040 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rncxb"] Dec 02 02:28:00 crc kubenswrapper[4884]: I1202 02:28:00.711382 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:00 crc kubenswrapper[4884]: I1202 02:28:00.729579 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rncxb"] Dec 02 02:28:00 crc kubenswrapper[4884]: I1202 02:28:00.827421 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cbaf926-1055-488d-aef1-e94a5db387a5-utilities\") pod \"certified-operators-rncxb\" (UID: \"2cbaf926-1055-488d-aef1-e94a5db387a5\") " pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:00 crc kubenswrapper[4884]: I1202 02:28:00.827722 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjvml\" (UniqueName: \"kubernetes.io/projected/2cbaf926-1055-488d-aef1-e94a5db387a5-kube-api-access-sjvml\") pod \"certified-operators-rncxb\" (UID: \"2cbaf926-1055-488d-aef1-e94a5db387a5\") " pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:00 crc kubenswrapper[4884]: I1202 02:28:00.827764 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cbaf926-1055-488d-aef1-e94a5db387a5-catalog-content\") pod \"certified-operators-rncxb\" (UID: \"2cbaf926-1055-488d-aef1-e94a5db387a5\") " pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:00 crc kubenswrapper[4884]: I1202 02:28:00.930004 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cbaf926-1055-488d-aef1-e94a5db387a5-utilities\") pod \"certified-operators-rncxb\" (UID: \"2cbaf926-1055-488d-aef1-e94a5db387a5\") " pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:00 crc kubenswrapper[4884]: I1202 02:28:00.930137 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjvml\" (UniqueName: \"kubernetes.io/projected/2cbaf926-1055-488d-aef1-e94a5db387a5-kube-api-access-sjvml\") pod \"certified-operators-rncxb\" (UID: \"2cbaf926-1055-488d-aef1-e94a5db387a5\") " pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:00 crc kubenswrapper[4884]: I1202 02:28:00.930189 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cbaf926-1055-488d-aef1-e94a5db387a5-catalog-content\") pod \"certified-operators-rncxb\" (UID: \"2cbaf926-1055-488d-aef1-e94a5db387a5\") " pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:00 crc kubenswrapper[4884]: I1202 02:28:00.930713 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cbaf926-1055-488d-aef1-e94a5db387a5-catalog-content\") pod \"certified-operators-rncxb\" (UID: \"2cbaf926-1055-488d-aef1-e94a5db387a5\") " pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:00 crc kubenswrapper[4884]: I1202 02:28:00.930942 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cbaf926-1055-488d-aef1-e94a5db387a5-utilities\") pod \"certified-operators-rncxb\" (UID: \"2cbaf926-1055-488d-aef1-e94a5db387a5\") " pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:00 crc kubenswrapper[4884]: I1202 02:28:00.951717 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjvml\" (UniqueName: \"kubernetes.io/projected/2cbaf926-1055-488d-aef1-e94a5db387a5-kube-api-access-sjvml\") pod \"certified-operators-rncxb\" (UID: \"2cbaf926-1055-488d-aef1-e94a5db387a5\") " pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:01 crc kubenswrapper[4884]: I1202 02:28:01.042702 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:01 crc kubenswrapper[4884]: I1202 02:28:01.579324 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rncxb"] Dec 02 02:28:02 crc kubenswrapper[4884]: I1202 02:28:02.021365 4884 generic.go:334] "Generic (PLEG): container finished" podID="2cbaf926-1055-488d-aef1-e94a5db387a5" containerID="5738bf87b3dad09517349f902fe3c0b6bb8815ec858719e1b23c58f9e68b511d" exitCode=0 Dec 02 02:28:02 crc kubenswrapper[4884]: I1202 02:28:02.021414 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rncxb" event={"ID":"2cbaf926-1055-488d-aef1-e94a5db387a5","Type":"ContainerDied","Data":"5738bf87b3dad09517349f902fe3c0b6bb8815ec858719e1b23c58f9e68b511d"} Dec 02 02:28:02 crc kubenswrapper[4884]: I1202 02:28:02.021721 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rncxb" event={"ID":"2cbaf926-1055-488d-aef1-e94a5db387a5","Type":"ContainerStarted","Data":"7f885ce91baf36141f925123afc83c3aea14e5e97cf09e6661784278ea12ff1f"} Dec 02 02:28:02 crc kubenswrapper[4884]: I1202 02:28:02.023699 4884 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 02:28:03 crc kubenswrapper[4884]: I1202 02:28:03.036090 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rncxb" event={"ID":"2cbaf926-1055-488d-aef1-e94a5db387a5","Type":"ContainerStarted","Data":"a7a07c253f35dce66ed10c9620ef839d97de06e5ff26a54f0e818e2071105d7f"} Dec 02 02:28:04 crc kubenswrapper[4884]: I1202 02:28:04.052790 4884 generic.go:334] "Generic (PLEG): container finished" podID="2cbaf926-1055-488d-aef1-e94a5db387a5" containerID="a7a07c253f35dce66ed10c9620ef839d97de06e5ff26a54f0e818e2071105d7f" exitCode=0 Dec 02 02:28:04 crc kubenswrapper[4884]: I1202 02:28:04.052923 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rncxb" event={"ID":"2cbaf926-1055-488d-aef1-e94a5db387a5","Type":"ContainerDied","Data":"a7a07c253f35dce66ed10c9620ef839d97de06e5ff26a54f0e818e2071105d7f"} Dec 02 02:28:04 crc kubenswrapper[4884]: I1202 02:28:04.881253 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-759q5"] Dec 02 02:28:04 crc kubenswrapper[4884]: I1202 02:28:04.884251 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:04 crc kubenswrapper[4884]: I1202 02:28:04.904165 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-759q5"] Dec 02 02:28:05 crc kubenswrapper[4884]: I1202 02:28:05.033164 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91d83142-5fea-4042-8b25-b8752fda6619-utilities\") pod \"redhat-operators-759q5\" (UID: \"91d83142-5fea-4042-8b25-b8752fda6619\") " pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:05 crc kubenswrapper[4884]: I1202 02:28:05.033241 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrn4n\" (UniqueName: \"kubernetes.io/projected/91d83142-5fea-4042-8b25-b8752fda6619-kube-api-access-rrn4n\") pod \"redhat-operators-759q5\" (UID: \"91d83142-5fea-4042-8b25-b8752fda6619\") " pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:05 crc kubenswrapper[4884]: I1202 02:28:05.033285 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91d83142-5fea-4042-8b25-b8752fda6619-catalog-content\") pod \"redhat-operators-759q5\" (UID: \"91d83142-5fea-4042-8b25-b8752fda6619\") " pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:05 crc kubenswrapper[4884]: I1202 02:28:05.065417 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rncxb" event={"ID":"2cbaf926-1055-488d-aef1-e94a5db387a5","Type":"ContainerStarted","Data":"33b91e95000dcbf4e5d079058675e60e23d151089fa4db32822e2c3391849d28"} Dec 02 02:28:05 crc kubenswrapper[4884]: I1202 02:28:05.092734 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rncxb" podStartSLOduration=2.573323936 podStartE2EDuration="5.092718443s" podCreationTimestamp="2025-12-02 02:28:00 +0000 UTC" firstStartedPulling="2025-12-02 02:28:02.023329708 +0000 UTC m=+3098.699166632" lastFinishedPulling="2025-12-02 02:28:04.542724245 +0000 UTC m=+3101.218561139" observedRunningTime="2025-12-02 02:28:05.084603067 +0000 UTC m=+3101.760439971" watchObservedRunningTime="2025-12-02 02:28:05.092718443 +0000 UTC m=+3101.768555327" Dec 02 02:28:05 crc kubenswrapper[4884]: I1202 02:28:05.135106 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91d83142-5fea-4042-8b25-b8752fda6619-utilities\") pod \"redhat-operators-759q5\" (UID: \"91d83142-5fea-4042-8b25-b8752fda6619\") " pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:05 crc kubenswrapper[4884]: I1202 02:28:05.136656 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91d83142-5fea-4042-8b25-b8752fda6619-utilities\") pod \"redhat-operators-759q5\" (UID: \"91d83142-5fea-4042-8b25-b8752fda6619\") " pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:05 crc kubenswrapper[4884]: I1202 02:28:05.136904 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrn4n\" (UniqueName: \"kubernetes.io/projected/91d83142-5fea-4042-8b25-b8752fda6619-kube-api-access-rrn4n\") pod \"redhat-operators-759q5\" (UID: \"91d83142-5fea-4042-8b25-b8752fda6619\") " pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:05 crc kubenswrapper[4884]: I1202 02:28:05.137057 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91d83142-5fea-4042-8b25-b8752fda6619-catalog-content\") pod \"redhat-operators-759q5\" (UID: \"91d83142-5fea-4042-8b25-b8752fda6619\") " pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:05 crc kubenswrapper[4884]: I1202 02:28:05.138628 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91d83142-5fea-4042-8b25-b8752fda6619-catalog-content\") pod \"redhat-operators-759q5\" (UID: \"91d83142-5fea-4042-8b25-b8752fda6619\") " pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:05 crc kubenswrapper[4884]: I1202 02:28:05.158665 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrn4n\" (UniqueName: \"kubernetes.io/projected/91d83142-5fea-4042-8b25-b8752fda6619-kube-api-access-rrn4n\") pod \"redhat-operators-759q5\" (UID: \"91d83142-5fea-4042-8b25-b8752fda6619\") " pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:05 crc kubenswrapper[4884]: I1202 02:28:05.247673 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:05 crc kubenswrapper[4884]: I1202 02:28:05.703195 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-759q5"] Dec 02 02:28:06 crc kubenswrapper[4884]: I1202 02:28:06.074410 4884 generic.go:334] "Generic (PLEG): container finished" podID="91d83142-5fea-4042-8b25-b8752fda6619" containerID="0b22eae7cc4788001199ffd095f2da7fe55f07edb6897a1af3e72f81dac3892f" exitCode=0 Dec 02 02:28:06 crc kubenswrapper[4884]: I1202 02:28:06.074466 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-759q5" event={"ID":"91d83142-5fea-4042-8b25-b8752fda6619","Type":"ContainerDied","Data":"0b22eae7cc4788001199ffd095f2da7fe55f07edb6897a1af3e72f81dac3892f"} Dec 02 02:28:06 crc kubenswrapper[4884]: I1202 02:28:06.075682 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-759q5" event={"ID":"91d83142-5fea-4042-8b25-b8752fda6619","Type":"ContainerStarted","Data":"761e42b1d5d8ecc2655f6d57e3f1ffb9c0806eddba39f667b036c4ffa4696a85"} Dec 02 02:28:08 crc kubenswrapper[4884]: I1202 02:28:08.103645 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-759q5" event={"ID":"91d83142-5fea-4042-8b25-b8752fda6619","Type":"ContainerStarted","Data":"31c98a006506c58839eeae3fb2fe6b8b550a10e61cc8c6a7079b99ec00ae2839"} Dec 02 02:28:10 crc kubenswrapper[4884]: I1202 02:28:10.131286 4884 generic.go:334] "Generic (PLEG): container finished" podID="91d83142-5fea-4042-8b25-b8752fda6619" containerID="31c98a006506c58839eeae3fb2fe6b8b550a10e61cc8c6a7079b99ec00ae2839" exitCode=0 Dec 02 02:28:10 crc kubenswrapper[4884]: I1202 02:28:10.131340 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-759q5" event={"ID":"91d83142-5fea-4042-8b25-b8752fda6619","Type":"ContainerDied","Data":"31c98a006506c58839eeae3fb2fe6b8b550a10e61cc8c6a7079b99ec00ae2839"} Dec 02 02:28:11 crc kubenswrapper[4884]: I1202 02:28:11.043439 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:11 crc kubenswrapper[4884]: I1202 02:28:11.043879 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:11 crc kubenswrapper[4884]: I1202 02:28:11.124352 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:11 crc kubenswrapper[4884]: I1202 02:28:11.201783 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:12 crc kubenswrapper[4884]: I1202 02:28:12.063649 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rncxb"] Dec 02 02:28:12 crc kubenswrapper[4884]: I1202 02:28:12.160992 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-759q5" event={"ID":"91d83142-5fea-4042-8b25-b8752fda6619","Type":"ContainerStarted","Data":"dfe04ffcecb159527583387ffa8faa1ac5bd1a8b01dfe3ca125dd58a9b8fa8da"} Dec 02 02:28:12 crc kubenswrapper[4884]: I1202 02:28:12.186710 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-759q5" podStartSLOduration=3.339724785 podStartE2EDuration="8.18668144s" podCreationTimestamp="2025-12-02 02:28:04 +0000 UTC" firstStartedPulling="2025-12-02 02:28:06.075867005 +0000 UTC m=+3102.751703899" lastFinishedPulling="2025-12-02 02:28:10.92282366 +0000 UTC m=+3107.598660554" observedRunningTime="2025-12-02 02:28:12.184502997 +0000 UTC m=+3108.860339931" watchObservedRunningTime="2025-12-02 02:28:12.18668144 +0000 UTC m=+3108.862518364" Dec 02 02:28:13 crc kubenswrapper[4884]: I1202 02:28:13.170335 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rncxb" podUID="2cbaf926-1055-488d-aef1-e94a5db387a5" containerName="registry-server" containerID="cri-o://33b91e95000dcbf4e5d079058675e60e23d151089fa4db32822e2c3391849d28" gracePeriod=2 Dec 02 02:28:13 crc kubenswrapper[4884]: I1202 02:28:13.644589 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:13 crc kubenswrapper[4884]: I1202 02:28:13.827883 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cbaf926-1055-488d-aef1-e94a5db387a5-utilities\") pod \"2cbaf926-1055-488d-aef1-e94a5db387a5\" (UID: \"2cbaf926-1055-488d-aef1-e94a5db387a5\") " Dec 02 02:28:13 crc kubenswrapper[4884]: I1202 02:28:13.828093 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cbaf926-1055-488d-aef1-e94a5db387a5-catalog-content\") pod \"2cbaf926-1055-488d-aef1-e94a5db387a5\" (UID: \"2cbaf926-1055-488d-aef1-e94a5db387a5\") " Dec 02 02:28:13 crc kubenswrapper[4884]: I1202 02:28:13.828177 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjvml\" (UniqueName: \"kubernetes.io/projected/2cbaf926-1055-488d-aef1-e94a5db387a5-kube-api-access-sjvml\") pod \"2cbaf926-1055-488d-aef1-e94a5db387a5\" (UID: \"2cbaf926-1055-488d-aef1-e94a5db387a5\") " Dec 02 02:28:13 crc kubenswrapper[4884]: I1202 02:28:13.828811 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cbaf926-1055-488d-aef1-e94a5db387a5-utilities" (OuterVolumeSpecName: "utilities") pod "2cbaf926-1055-488d-aef1-e94a5db387a5" (UID: "2cbaf926-1055-488d-aef1-e94a5db387a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:28:13 crc kubenswrapper[4884]: I1202 02:28:13.846931 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cbaf926-1055-488d-aef1-e94a5db387a5-kube-api-access-sjvml" (OuterVolumeSpecName: "kube-api-access-sjvml") pod "2cbaf926-1055-488d-aef1-e94a5db387a5" (UID: "2cbaf926-1055-488d-aef1-e94a5db387a5"). InnerVolumeSpecName "kube-api-access-sjvml". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:28:13 crc kubenswrapper[4884]: I1202 02:28:13.898508 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cbaf926-1055-488d-aef1-e94a5db387a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2cbaf926-1055-488d-aef1-e94a5db387a5" (UID: "2cbaf926-1055-488d-aef1-e94a5db387a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:28:13 crc kubenswrapper[4884]: I1202 02:28:13.931320 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjvml\" (UniqueName: \"kubernetes.io/projected/2cbaf926-1055-488d-aef1-e94a5db387a5-kube-api-access-sjvml\") on node \"crc\" DevicePath \"\"" Dec 02 02:28:13 crc kubenswrapper[4884]: I1202 02:28:13.931360 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cbaf926-1055-488d-aef1-e94a5db387a5-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:28:13 crc kubenswrapper[4884]: I1202 02:28:13.931376 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cbaf926-1055-488d-aef1-e94a5db387a5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:28:14 crc kubenswrapper[4884]: I1202 02:28:14.183076 4884 generic.go:334] "Generic (PLEG): container finished" podID="2cbaf926-1055-488d-aef1-e94a5db387a5" containerID="33b91e95000dcbf4e5d079058675e60e23d151089fa4db32822e2c3391849d28" exitCode=0 Dec 02 02:28:14 crc kubenswrapper[4884]: I1202 02:28:14.183421 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rncxb" event={"ID":"2cbaf926-1055-488d-aef1-e94a5db387a5","Type":"ContainerDied","Data":"33b91e95000dcbf4e5d079058675e60e23d151089fa4db32822e2c3391849d28"} Dec 02 02:28:14 crc kubenswrapper[4884]: I1202 02:28:14.183461 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rncxb" event={"ID":"2cbaf926-1055-488d-aef1-e94a5db387a5","Type":"ContainerDied","Data":"7f885ce91baf36141f925123afc83c3aea14e5e97cf09e6661784278ea12ff1f"} Dec 02 02:28:14 crc kubenswrapper[4884]: I1202 02:28:14.183491 4884 scope.go:117] "RemoveContainer" containerID="33b91e95000dcbf4e5d079058675e60e23d151089fa4db32822e2c3391849d28" Dec 02 02:28:14 crc kubenswrapper[4884]: I1202 02:28:14.183699 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rncxb" Dec 02 02:28:14 crc kubenswrapper[4884]: I1202 02:28:14.214047 4884 scope.go:117] "RemoveContainer" containerID="a7a07c253f35dce66ed10c9620ef839d97de06e5ff26a54f0e818e2071105d7f" Dec 02 02:28:14 crc kubenswrapper[4884]: I1202 02:28:14.244438 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rncxb"] Dec 02 02:28:14 crc kubenswrapper[4884]: I1202 02:28:14.249973 4884 scope.go:117] "RemoveContainer" containerID="5738bf87b3dad09517349f902fe3c0b6bb8815ec858719e1b23c58f9e68b511d" Dec 02 02:28:14 crc kubenswrapper[4884]: I1202 02:28:14.259547 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rncxb"] Dec 02 02:28:14 crc kubenswrapper[4884]: I1202 02:28:14.314949 4884 scope.go:117] "RemoveContainer" containerID="33b91e95000dcbf4e5d079058675e60e23d151089fa4db32822e2c3391849d28" Dec 02 02:28:14 crc kubenswrapper[4884]: E1202 02:28:14.315490 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33b91e95000dcbf4e5d079058675e60e23d151089fa4db32822e2c3391849d28\": container with ID starting with 33b91e95000dcbf4e5d079058675e60e23d151089fa4db32822e2c3391849d28 not found: ID does not exist" containerID="33b91e95000dcbf4e5d079058675e60e23d151089fa4db32822e2c3391849d28" Dec 02 02:28:14 crc kubenswrapper[4884]: I1202 02:28:14.315537 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33b91e95000dcbf4e5d079058675e60e23d151089fa4db32822e2c3391849d28"} err="failed to get container status \"33b91e95000dcbf4e5d079058675e60e23d151089fa4db32822e2c3391849d28\": rpc error: code = NotFound desc = could not find container \"33b91e95000dcbf4e5d079058675e60e23d151089fa4db32822e2c3391849d28\": container with ID starting with 33b91e95000dcbf4e5d079058675e60e23d151089fa4db32822e2c3391849d28 not found: ID does not exist" Dec 02 02:28:14 crc kubenswrapper[4884]: I1202 02:28:14.315566 4884 scope.go:117] "RemoveContainer" containerID="a7a07c253f35dce66ed10c9620ef839d97de06e5ff26a54f0e818e2071105d7f" Dec 02 02:28:14 crc kubenswrapper[4884]: E1202 02:28:14.316049 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7a07c253f35dce66ed10c9620ef839d97de06e5ff26a54f0e818e2071105d7f\": container with ID starting with a7a07c253f35dce66ed10c9620ef839d97de06e5ff26a54f0e818e2071105d7f not found: ID does not exist" containerID="a7a07c253f35dce66ed10c9620ef839d97de06e5ff26a54f0e818e2071105d7f" Dec 02 02:28:14 crc kubenswrapper[4884]: I1202 02:28:14.316091 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7a07c253f35dce66ed10c9620ef839d97de06e5ff26a54f0e818e2071105d7f"} err="failed to get container status \"a7a07c253f35dce66ed10c9620ef839d97de06e5ff26a54f0e818e2071105d7f\": rpc error: code = NotFound desc = could not find container \"a7a07c253f35dce66ed10c9620ef839d97de06e5ff26a54f0e818e2071105d7f\": container with ID starting with a7a07c253f35dce66ed10c9620ef839d97de06e5ff26a54f0e818e2071105d7f not found: ID does not exist" Dec 02 02:28:14 crc kubenswrapper[4884]: I1202 02:28:14.316120 4884 scope.go:117] "RemoveContainer" containerID="5738bf87b3dad09517349f902fe3c0b6bb8815ec858719e1b23c58f9e68b511d" Dec 02 02:28:14 crc kubenswrapper[4884]: E1202 02:28:14.316458 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5738bf87b3dad09517349f902fe3c0b6bb8815ec858719e1b23c58f9e68b511d\": container with ID starting with 5738bf87b3dad09517349f902fe3c0b6bb8815ec858719e1b23c58f9e68b511d not found: ID does not exist" containerID="5738bf87b3dad09517349f902fe3c0b6bb8815ec858719e1b23c58f9e68b511d" Dec 02 02:28:14 crc kubenswrapper[4884]: I1202 02:28:14.316485 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5738bf87b3dad09517349f902fe3c0b6bb8815ec858719e1b23c58f9e68b511d"} err="failed to get container status \"5738bf87b3dad09517349f902fe3c0b6bb8815ec858719e1b23c58f9e68b511d\": rpc error: code = NotFound desc = could not find container \"5738bf87b3dad09517349f902fe3c0b6bb8815ec858719e1b23c58f9e68b511d\": container with ID starting with 5738bf87b3dad09517349f902fe3c0b6bb8815ec858719e1b23c58f9e68b511d not found: ID does not exist" Dec 02 02:28:15 crc kubenswrapper[4884]: I1202 02:28:15.249065 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:15 crc kubenswrapper[4884]: I1202 02:28:15.249147 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:15 crc kubenswrapper[4884]: I1202 02:28:15.634075 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cbaf926-1055-488d-aef1-e94a5db387a5" path="/var/lib/kubelet/pods/2cbaf926-1055-488d-aef1-e94a5db387a5/volumes" Dec 02 02:28:16 crc kubenswrapper[4884]: I1202 02:28:16.309829 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-759q5" podUID="91d83142-5fea-4042-8b25-b8752fda6619" containerName="registry-server" probeResult="failure" output=< Dec 02 02:28:16 crc kubenswrapper[4884]: timeout: failed to connect service ":50051" within 1s Dec 02 02:28:16 crc kubenswrapper[4884]: > Dec 02 02:28:16 crc kubenswrapper[4884]: I1202 02:28:16.971561 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:28:16 crc kubenswrapper[4884]: I1202 02:28:16.971645 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:28:16 crc kubenswrapper[4884]: I1202 02:28:16.971706 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 02:28:16 crc kubenswrapper[4884]: I1202 02:28:16.972636 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 02:28:16 crc kubenswrapper[4884]: I1202 02:28:16.972715 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" gracePeriod=600 Dec 02 02:28:17 crc kubenswrapper[4884]: E1202 02:28:17.106710 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:28:17 crc kubenswrapper[4884]: E1202 02:28:17.175131 4884 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9385a5c9_a156_460a_8ca1_9b342ab835c9.slice/crio-conmon-bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9385a5c9_a156_460a_8ca1_9b342ab835c9.slice/crio-bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9.scope\": RecentStats: unable to find data in memory cache]" Dec 02 02:28:17 crc kubenswrapper[4884]: I1202 02:28:17.221229 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" exitCode=0 Dec 02 02:28:17 crc kubenswrapper[4884]: I1202 02:28:17.221275 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9"} Dec 02 02:28:17 crc kubenswrapper[4884]: I1202 02:28:17.221309 4884 scope.go:117] "RemoveContainer" containerID="83598bb7142be85f0e3373487befb0fe500a455b9b41c9b4af785be9d73c5ae8" Dec 02 02:28:17 crc kubenswrapper[4884]: I1202 02:28:17.221842 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:28:17 crc kubenswrapper[4884]: E1202 02:28:17.222189 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:28:25 crc kubenswrapper[4884]: I1202 02:28:25.327488 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:25 crc kubenswrapper[4884]: I1202 02:28:25.405112 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:25 crc kubenswrapper[4884]: I1202 02:28:25.578282 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-759q5"] Dec 02 02:28:26 crc kubenswrapper[4884]: I1202 02:28:26.367240 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-759q5" podUID="91d83142-5fea-4042-8b25-b8752fda6619" containerName="registry-server" containerID="cri-o://dfe04ffcecb159527583387ffa8faa1ac5bd1a8b01dfe3ca125dd58a9b8fa8da" gracePeriod=2 Dec 02 02:28:26 crc kubenswrapper[4884]: I1202 02:28:26.827159 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:26 crc kubenswrapper[4884]: I1202 02:28:26.933757 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91d83142-5fea-4042-8b25-b8752fda6619-utilities\") pod \"91d83142-5fea-4042-8b25-b8752fda6619\" (UID: \"91d83142-5fea-4042-8b25-b8752fda6619\") " Dec 02 02:28:26 crc kubenswrapper[4884]: I1202 02:28:26.933837 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrn4n\" (UniqueName: \"kubernetes.io/projected/91d83142-5fea-4042-8b25-b8752fda6619-kube-api-access-rrn4n\") pod \"91d83142-5fea-4042-8b25-b8752fda6619\" (UID: \"91d83142-5fea-4042-8b25-b8752fda6619\") " Dec 02 02:28:26 crc kubenswrapper[4884]: I1202 02:28:26.934012 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91d83142-5fea-4042-8b25-b8752fda6619-catalog-content\") pod \"91d83142-5fea-4042-8b25-b8752fda6619\" (UID: \"91d83142-5fea-4042-8b25-b8752fda6619\") " Dec 02 02:28:26 crc kubenswrapper[4884]: I1202 02:28:26.934542 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91d83142-5fea-4042-8b25-b8752fda6619-utilities" (OuterVolumeSpecName: "utilities") pod "91d83142-5fea-4042-8b25-b8752fda6619" (UID: "91d83142-5fea-4042-8b25-b8752fda6619"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:28:26 crc kubenswrapper[4884]: I1202 02:28:26.934924 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91d83142-5fea-4042-8b25-b8752fda6619-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:28:26 crc kubenswrapper[4884]: I1202 02:28:26.939238 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91d83142-5fea-4042-8b25-b8752fda6619-kube-api-access-rrn4n" (OuterVolumeSpecName: "kube-api-access-rrn4n") pod "91d83142-5fea-4042-8b25-b8752fda6619" (UID: "91d83142-5fea-4042-8b25-b8752fda6619"). InnerVolumeSpecName "kube-api-access-rrn4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.036462 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrn4n\" (UniqueName: \"kubernetes.io/projected/91d83142-5fea-4042-8b25-b8752fda6619-kube-api-access-rrn4n\") on node \"crc\" DevicePath \"\"" Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.059917 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91d83142-5fea-4042-8b25-b8752fda6619-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91d83142-5fea-4042-8b25-b8752fda6619" (UID: "91d83142-5fea-4042-8b25-b8752fda6619"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.137919 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91d83142-5fea-4042-8b25-b8752fda6619-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.382752 4884 generic.go:334] "Generic (PLEG): container finished" podID="91d83142-5fea-4042-8b25-b8752fda6619" containerID="dfe04ffcecb159527583387ffa8faa1ac5bd1a8b01dfe3ca125dd58a9b8fa8da" exitCode=0 Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.382833 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-759q5" event={"ID":"91d83142-5fea-4042-8b25-b8752fda6619","Type":"ContainerDied","Data":"dfe04ffcecb159527583387ffa8faa1ac5bd1a8b01dfe3ca125dd58a9b8fa8da"} Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.382866 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-759q5" event={"ID":"91d83142-5fea-4042-8b25-b8752fda6619","Type":"ContainerDied","Data":"761e42b1d5d8ecc2655f6d57e3f1ffb9c0806eddba39f667b036c4ffa4696a85"} Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.382885 4884 scope.go:117] "RemoveContainer" containerID="dfe04ffcecb159527583387ffa8faa1ac5bd1a8b01dfe3ca125dd58a9b8fa8da" Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.383013 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-759q5" Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.438599 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-759q5"] Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.439771 4884 scope.go:117] "RemoveContainer" containerID="31c98a006506c58839eeae3fb2fe6b8b550a10e61cc8c6a7079b99ec00ae2839" Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.450042 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-759q5"] Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.469897 4884 scope.go:117] "RemoveContainer" containerID="0b22eae7cc4788001199ffd095f2da7fe55f07edb6897a1af3e72f81dac3892f" Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.522446 4884 scope.go:117] "RemoveContainer" containerID="dfe04ffcecb159527583387ffa8faa1ac5bd1a8b01dfe3ca125dd58a9b8fa8da" Dec 02 02:28:27 crc kubenswrapper[4884]: E1202 02:28:27.523213 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfe04ffcecb159527583387ffa8faa1ac5bd1a8b01dfe3ca125dd58a9b8fa8da\": container with ID starting with dfe04ffcecb159527583387ffa8faa1ac5bd1a8b01dfe3ca125dd58a9b8fa8da not found: ID does not exist" containerID="dfe04ffcecb159527583387ffa8faa1ac5bd1a8b01dfe3ca125dd58a9b8fa8da" Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.523244 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfe04ffcecb159527583387ffa8faa1ac5bd1a8b01dfe3ca125dd58a9b8fa8da"} err="failed to get container status \"dfe04ffcecb159527583387ffa8faa1ac5bd1a8b01dfe3ca125dd58a9b8fa8da\": rpc error: code = NotFound desc = could not find container \"dfe04ffcecb159527583387ffa8faa1ac5bd1a8b01dfe3ca125dd58a9b8fa8da\": container with ID starting with dfe04ffcecb159527583387ffa8faa1ac5bd1a8b01dfe3ca125dd58a9b8fa8da not found: ID does not exist" Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.523272 4884 scope.go:117] "RemoveContainer" containerID="31c98a006506c58839eeae3fb2fe6b8b550a10e61cc8c6a7079b99ec00ae2839" Dec 02 02:28:27 crc kubenswrapper[4884]: E1202 02:28:27.523849 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31c98a006506c58839eeae3fb2fe6b8b550a10e61cc8c6a7079b99ec00ae2839\": container with ID starting with 31c98a006506c58839eeae3fb2fe6b8b550a10e61cc8c6a7079b99ec00ae2839 not found: ID does not exist" containerID="31c98a006506c58839eeae3fb2fe6b8b550a10e61cc8c6a7079b99ec00ae2839" Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.523880 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31c98a006506c58839eeae3fb2fe6b8b550a10e61cc8c6a7079b99ec00ae2839"} err="failed to get container status \"31c98a006506c58839eeae3fb2fe6b8b550a10e61cc8c6a7079b99ec00ae2839\": rpc error: code = NotFound desc = could not find container \"31c98a006506c58839eeae3fb2fe6b8b550a10e61cc8c6a7079b99ec00ae2839\": container with ID starting with 31c98a006506c58839eeae3fb2fe6b8b550a10e61cc8c6a7079b99ec00ae2839 not found: ID does not exist" Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.523898 4884 scope.go:117] "RemoveContainer" containerID="0b22eae7cc4788001199ffd095f2da7fe55f07edb6897a1af3e72f81dac3892f" Dec 02 02:28:27 crc kubenswrapper[4884]: E1202 02:28:27.524201 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b22eae7cc4788001199ffd095f2da7fe55f07edb6897a1af3e72f81dac3892f\": container with ID starting with 0b22eae7cc4788001199ffd095f2da7fe55f07edb6897a1af3e72f81dac3892f not found: ID does not exist" containerID="0b22eae7cc4788001199ffd095f2da7fe55f07edb6897a1af3e72f81dac3892f" Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.524220 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b22eae7cc4788001199ffd095f2da7fe55f07edb6897a1af3e72f81dac3892f"} err="failed to get container status \"0b22eae7cc4788001199ffd095f2da7fe55f07edb6897a1af3e72f81dac3892f\": rpc error: code = NotFound desc = could not find container \"0b22eae7cc4788001199ffd095f2da7fe55f07edb6897a1af3e72f81dac3892f\": container with ID starting with 0b22eae7cc4788001199ffd095f2da7fe55f07edb6897a1af3e72f81dac3892f not found: ID does not exist" Dec 02 02:28:27 crc kubenswrapper[4884]: I1202 02:28:27.625456 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91d83142-5fea-4042-8b25-b8752fda6619" path="/var/lib/kubelet/pods/91d83142-5fea-4042-8b25-b8752fda6619/volumes" Dec 02 02:28:28 crc kubenswrapper[4884]: I1202 02:28:28.614722 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:28:28 crc kubenswrapper[4884]: E1202 02:28:28.615258 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:28:39 crc kubenswrapper[4884]: I1202 02:28:39.615824 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:28:39 crc kubenswrapper[4884]: E1202 02:28:39.617069 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:28:50 crc kubenswrapper[4884]: I1202 02:28:50.616284 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:28:50 crc kubenswrapper[4884]: E1202 02:28:50.617670 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:29:02 crc kubenswrapper[4884]: I1202 02:29:02.614593 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:29:02 crc kubenswrapper[4884]: E1202 02:29:02.615892 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:29:16 crc kubenswrapper[4884]: I1202 02:29:16.614454 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:29:16 crc kubenswrapper[4884]: E1202 02:29:16.615302 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:29:28 crc kubenswrapper[4884]: I1202 02:29:28.614671 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:29:28 crc kubenswrapper[4884]: E1202 02:29:28.615490 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:29:41 crc kubenswrapper[4884]: I1202 02:29:41.615441 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:29:41 crc kubenswrapper[4884]: E1202 02:29:41.616869 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:29:53 crc kubenswrapper[4884]: I1202 02:29:53.639891 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:29:53 crc kubenswrapper[4884]: E1202 02:29:53.640768 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.161277 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx"] Dec 02 02:30:00 crc kubenswrapper[4884]: E1202 02:30:00.162175 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cbaf926-1055-488d-aef1-e94a5db387a5" containerName="extract-utilities" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.162191 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cbaf926-1055-488d-aef1-e94a5db387a5" containerName="extract-utilities" Dec 02 02:30:00 crc kubenswrapper[4884]: E1202 02:30:00.162205 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cbaf926-1055-488d-aef1-e94a5db387a5" containerName="registry-server" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.162214 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cbaf926-1055-488d-aef1-e94a5db387a5" containerName="registry-server" Dec 02 02:30:00 crc kubenswrapper[4884]: E1202 02:30:00.162240 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cbaf926-1055-488d-aef1-e94a5db387a5" containerName="extract-content" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.162248 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cbaf926-1055-488d-aef1-e94a5db387a5" containerName="extract-content" Dec 02 02:30:00 crc kubenswrapper[4884]: E1202 02:30:00.162273 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d83142-5fea-4042-8b25-b8752fda6619" containerName="registry-server" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.162281 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d83142-5fea-4042-8b25-b8752fda6619" containerName="registry-server" Dec 02 02:30:00 crc kubenswrapper[4884]: E1202 02:30:00.162299 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d83142-5fea-4042-8b25-b8752fda6619" containerName="extract-content" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.162307 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d83142-5fea-4042-8b25-b8752fda6619" containerName="extract-content" Dec 02 02:30:00 crc kubenswrapper[4884]: E1202 02:30:00.162319 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91d83142-5fea-4042-8b25-b8752fda6619" containerName="extract-utilities" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.162326 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="91d83142-5fea-4042-8b25-b8752fda6619" containerName="extract-utilities" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.162572 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cbaf926-1055-488d-aef1-e94a5db387a5" containerName="registry-server" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.162594 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="91d83142-5fea-4042-8b25-b8752fda6619" containerName="registry-server" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.163361 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.171762 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx"] Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.179136 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.179297 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.295906 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-secret-volume\") pod \"collect-profiles-29410710-7lppx\" (UID: \"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.296064 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gm87\" (UniqueName: \"kubernetes.io/projected/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-kube-api-access-4gm87\") pod \"collect-profiles-29410710-7lppx\" (UID: \"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.296149 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-config-volume\") pod \"collect-profiles-29410710-7lppx\" (UID: \"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.397559 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-secret-volume\") pod \"collect-profiles-29410710-7lppx\" (UID: \"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.397723 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gm87\" (UniqueName: \"kubernetes.io/projected/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-kube-api-access-4gm87\") pod \"collect-profiles-29410710-7lppx\" (UID: \"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.397855 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-config-volume\") pod \"collect-profiles-29410710-7lppx\" (UID: \"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.399571 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-config-volume\") pod \"collect-profiles-29410710-7lppx\" (UID: \"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.406052 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-secret-volume\") pod \"collect-profiles-29410710-7lppx\" (UID: \"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.417601 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gm87\" (UniqueName: \"kubernetes.io/projected/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-kube-api-access-4gm87\") pod \"collect-profiles-29410710-7lppx\" (UID: \"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.496984 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" Dec 02 02:30:00 crc kubenswrapper[4884]: I1202 02:30:00.985332 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx"] Dec 02 02:30:01 crc kubenswrapper[4884]: W1202 02:30:00.999610 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podddf3d6bf_05f6_4097_9b4f_c7d5df4cffd5.slice/crio-c1299b7d85ec59c4dba8921538fcf8056391ada9a5feccde3826226a2e1748aa WatchSource:0}: Error finding container c1299b7d85ec59c4dba8921538fcf8056391ada9a5feccde3826226a2e1748aa: Status 404 returned error can't find the container with id c1299b7d85ec59c4dba8921538fcf8056391ada9a5feccde3826226a2e1748aa Dec 02 02:30:01 crc kubenswrapper[4884]: I1202 02:30:01.229391 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" event={"ID":"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5","Type":"ContainerStarted","Data":"4f9e7fa26d396a1e88176af67d53c99f72b00a3e13625da16517f8f000adf4ce"} Dec 02 02:30:01 crc kubenswrapper[4884]: I1202 02:30:01.229702 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" event={"ID":"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5","Type":"ContainerStarted","Data":"c1299b7d85ec59c4dba8921538fcf8056391ada9a5feccde3826226a2e1748aa"} Dec 02 02:30:01 crc kubenswrapper[4884]: I1202 02:30:01.247172 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" podStartSLOduration=1.247148916 podStartE2EDuration="1.247148916s" podCreationTimestamp="2025-12-02 02:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 02:30:01.245011084 +0000 UTC m=+3217.920847978" watchObservedRunningTime="2025-12-02 02:30:01.247148916 +0000 UTC m=+3217.922985810" Dec 02 02:30:02 crc kubenswrapper[4884]: I1202 02:30:02.239284 4884 generic.go:334] "Generic (PLEG): container finished" podID="ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5" containerID="4f9e7fa26d396a1e88176af67d53c99f72b00a3e13625da16517f8f000adf4ce" exitCode=0 Dec 02 02:30:02 crc kubenswrapper[4884]: I1202 02:30:02.239406 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" event={"ID":"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5","Type":"ContainerDied","Data":"4f9e7fa26d396a1e88176af67d53c99f72b00a3e13625da16517f8f000adf4ce"} Dec 02 02:30:03 crc kubenswrapper[4884]: I1202 02:30:03.674179 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" Dec 02 02:30:03 crc kubenswrapper[4884]: I1202 02:30:03.870763 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-secret-volume\") pod \"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5\" (UID: \"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5\") " Dec 02 02:30:03 crc kubenswrapper[4884]: I1202 02:30:03.870885 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-config-volume\") pod \"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5\" (UID: \"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5\") " Dec 02 02:30:03 crc kubenswrapper[4884]: I1202 02:30:03.871152 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gm87\" (UniqueName: \"kubernetes.io/projected/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-kube-api-access-4gm87\") pod \"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5\" (UID: \"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5\") " Dec 02 02:30:03 crc kubenswrapper[4884]: I1202 02:30:03.873054 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-config-volume" (OuterVolumeSpecName: "config-volume") pod "ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5" (UID: "ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:30:03 crc kubenswrapper[4884]: I1202 02:30:03.879906 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-kube-api-access-4gm87" (OuterVolumeSpecName: "kube-api-access-4gm87") pod "ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5" (UID: "ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5"). InnerVolumeSpecName "kube-api-access-4gm87". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:30:03 crc kubenswrapper[4884]: I1202 02:30:03.880177 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5" (UID: "ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:30:03 crc kubenswrapper[4884]: I1202 02:30:03.973740 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gm87\" (UniqueName: \"kubernetes.io/projected/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-kube-api-access-4gm87\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:03 crc kubenswrapper[4884]: I1202 02:30:03.974016 4884 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:03 crc kubenswrapper[4884]: I1202 02:30:03.974225 4884 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:04 crc kubenswrapper[4884]: I1202 02:30:04.265078 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" event={"ID":"ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5","Type":"ContainerDied","Data":"c1299b7d85ec59c4dba8921538fcf8056391ada9a5feccde3826226a2e1748aa"} Dec 02 02:30:04 crc kubenswrapper[4884]: I1202 02:30:04.265676 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1299b7d85ec59c4dba8921538fcf8056391ada9a5feccde3826226a2e1748aa" Dec 02 02:30:04 crc kubenswrapper[4884]: I1202 02:30:04.265325 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx" Dec 02 02:30:04 crc kubenswrapper[4884]: I1202 02:30:04.334805 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz"] Dec 02 02:30:04 crc kubenswrapper[4884]: I1202 02:30:04.344873 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410665-fx5wz"] Dec 02 02:30:05 crc kubenswrapper[4884]: I1202 02:30:05.635087 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55b9ef1b-0207-4a5e-97e8-8c05843fb6b4" path="/var/lib/kubelet/pods/55b9ef1b-0207-4a5e-97e8-8c05843fb6b4/volumes" Dec 02 02:30:08 crc kubenswrapper[4884]: I1202 02:30:08.614770 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:30:08 crc kubenswrapper[4884]: E1202 02:30:08.615446 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:30:19 crc kubenswrapper[4884]: I1202 02:30:19.430381 4884 generic.go:334] "Generic (PLEG): container finished" podID="7098a534-9f0a-457d-a3d2-09249bd490ee" containerID="bf2bc2c42e09fd749876f3c7a24faad004820eebf2a3518f90aa1b5017b9a131" exitCode=0 Dec 02 02:30:19 crc kubenswrapper[4884]: I1202 02:30:19.430464 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" event={"ID":"7098a534-9f0a-457d-a3d2-09249bd490ee","Type":"ContainerDied","Data":"bf2bc2c42e09fd749876f3c7a24faad004820eebf2a3518f90aa1b5017b9a131"} Dec 02 02:30:20 crc kubenswrapper[4884]: I1202 02:30:20.923431 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.060428 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-1\") pod \"7098a534-9f0a-457d-a3d2-09249bd490ee\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.060503 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ssh-key\") pod \"7098a534-9f0a-457d-a3d2-09249bd490ee\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.060557 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-2\") pod \"7098a534-9f0a-457d-a3d2-09249bd490ee\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.060661 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx62g\" (UniqueName: \"kubernetes.io/projected/7098a534-9f0a-457d-a3d2-09249bd490ee-kube-api-access-jx62g\") pod \"7098a534-9f0a-457d-a3d2-09249bd490ee\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.060736 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-inventory\") pod \"7098a534-9f0a-457d-a3d2-09249bd490ee\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.060789 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-telemetry-combined-ca-bundle\") pod \"7098a534-9f0a-457d-a3d2-09249bd490ee\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.060918 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-0\") pod \"7098a534-9f0a-457d-a3d2-09249bd490ee\" (UID: \"7098a534-9f0a-457d-a3d2-09249bd490ee\") " Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.072005 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "7098a534-9f0a-457d-a3d2-09249bd490ee" (UID: "7098a534-9f0a-457d-a3d2-09249bd490ee"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.072160 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7098a534-9f0a-457d-a3d2-09249bd490ee-kube-api-access-jx62g" (OuterVolumeSpecName: "kube-api-access-jx62g") pod "7098a534-9f0a-457d-a3d2-09249bd490ee" (UID: "7098a534-9f0a-457d-a3d2-09249bd490ee"). InnerVolumeSpecName "kube-api-access-jx62g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.098244 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-inventory" (OuterVolumeSpecName: "inventory") pod "7098a534-9f0a-457d-a3d2-09249bd490ee" (UID: "7098a534-9f0a-457d-a3d2-09249bd490ee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.106212 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "7098a534-9f0a-457d-a3d2-09249bd490ee" (UID: "7098a534-9f0a-457d-a3d2-09249bd490ee"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.106631 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "7098a534-9f0a-457d-a3d2-09249bd490ee" (UID: "7098a534-9f0a-457d-a3d2-09249bd490ee"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.109673 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7098a534-9f0a-457d-a3d2-09249bd490ee" (UID: "7098a534-9f0a-457d-a3d2-09249bd490ee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.111901 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "7098a534-9f0a-457d-a3d2-09249bd490ee" (UID: "7098a534-9f0a-457d-a3d2-09249bd490ee"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.163670 4884 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-inventory\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.163706 4884 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.163717 4884 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.163726 4884 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.163761 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.163774 4884 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/7098a534-9f0a-457d-a3d2-09249bd490ee-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.163787 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx62g\" (UniqueName: \"kubernetes.io/projected/7098a534-9f0a-457d-a3d2-09249bd490ee-kube-api-access-jx62g\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.459111 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" event={"ID":"7098a534-9f0a-457d-a3d2-09249bd490ee","Type":"ContainerDied","Data":"4b28e939d456cd51cabb331a6c32828b973e9c1e626af1b11e9295c78bbe13b5"} Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.459156 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz" Dec 02 02:30:21 crc kubenswrapper[4884]: I1202 02:30:21.459168 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b28e939d456cd51cabb331a6c32828b973e9c1e626af1b11e9295c78bbe13b5" Dec 02 02:30:23 crc kubenswrapper[4884]: I1202 02:30:23.625870 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:30:23 crc kubenswrapper[4884]: E1202 02:30:23.628005 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:30:35 crc kubenswrapper[4884]: I1202 02:30:35.614378 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:30:35 crc kubenswrapper[4884]: E1202 02:30:35.615452 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:30:46 crc kubenswrapper[4884]: I1202 02:30:46.286303 4884 scope.go:117] "RemoveContainer" containerID="db9281b4fb77ff1a2b2d9ee00b0e9f952c72aa088e60ae0a985a21297aa1aeb0" Dec 02 02:30:48 crc kubenswrapper[4884]: I1202 02:30:48.615305 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:30:48 crc kubenswrapper[4884]: E1202 02:30:48.616299 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:30:58 crc kubenswrapper[4884]: I1202 02:30:58.338891 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 02 02:30:58 crc kubenswrapper[4884]: I1202 02:30:58.342133 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerName="prometheus" containerID="cri-o://ff2a72f07e42090c0d6e3254cf3b17dd9ef3b8f1b498858eb94632c5a0884931" gracePeriod=600 Dec 02 02:30:58 crc kubenswrapper[4884]: I1202 02:30:58.342364 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerName="thanos-sidecar" containerID="cri-o://6103d609ead7e74a2f2a809ab023c79ca78a2835c48a603c317ecbf6f95d38bd" gracePeriod=600 Dec 02 02:30:58 crc kubenswrapper[4884]: I1202 02:30:58.342436 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerName="config-reloader" containerID="cri-o://22cae3ab5f411ea5f30ea9e8ad3ff278d413825ae977a8ee143fd766e1cf651a" gracePeriod=600 Dec 02 02:30:58 crc kubenswrapper[4884]: I1202 02:30:58.957224 4884 generic.go:334] "Generic (PLEG): container finished" podID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerID="6103d609ead7e74a2f2a809ab023c79ca78a2835c48a603c317ecbf6f95d38bd" exitCode=0 Dec 02 02:30:58 crc kubenswrapper[4884]: I1202 02:30:58.957265 4884 generic.go:334] "Generic (PLEG): container finished" podID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerID="22cae3ab5f411ea5f30ea9e8ad3ff278d413825ae977a8ee143fd766e1cf651a" exitCode=0 Dec 02 02:30:58 crc kubenswrapper[4884]: I1202 02:30:58.957275 4884 generic.go:334] "Generic (PLEG): container finished" podID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerID="ff2a72f07e42090c0d6e3254cf3b17dd9ef3b8f1b498858eb94632c5a0884931" exitCode=0 Dec 02 02:30:58 crc kubenswrapper[4884]: I1202 02:30:58.957311 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aa5b4806-95dc-4496-81cd-d66b3ad23c13","Type":"ContainerDied","Data":"6103d609ead7e74a2f2a809ab023c79ca78a2835c48a603c317ecbf6f95d38bd"} Dec 02 02:30:58 crc kubenswrapper[4884]: I1202 02:30:58.957372 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aa5b4806-95dc-4496-81cd-d66b3ad23c13","Type":"ContainerDied","Data":"22cae3ab5f411ea5f30ea9e8ad3ff278d413825ae977a8ee143fd766e1cf651a"} Dec 02 02:30:58 crc kubenswrapper[4884]: I1202 02:30:58.957392 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aa5b4806-95dc-4496-81cd-d66b3ad23c13","Type":"ContainerDied","Data":"ff2a72f07e42090c0d6e3254cf3b17dd9ef3b8f1b498858eb94632c5a0884931"} Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.420604 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.542805 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aa5b4806-95dc-4496-81cd-d66b3ad23c13-config-out\") pod \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.542867 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-secret-combined-ca-bundle\") pod \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.542891 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-config\") pod \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.543093 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") pod \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.543181 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aa5b4806-95dc-4496-81cd-d66b3ad23c13-prometheus-metric-storage-rulefiles-0\") pod \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.543214 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config\") pod \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.543267 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-thanos-prometheus-http-client-file\") pod \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.543353 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.543389 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5ngs\" (UniqueName: \"kubernetes.io/projected/aa5b4806-95dc-4496-81cd-d66b3ad23c13-kube-api-access-r5ngs\") pod \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.543406 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.543444 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aa5b4806-95dc-4496-81cd-d66b3ad23c13-tls-assets\") pod \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\" (UID: \"aa5b4806-95dc-4496-81cd-d66b3ad23c13\") " Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.544341 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa5b4806-95dc-4496-81cd-d66b3ad23c13-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "aa5b4806-95dc-4496-81cd-d66b3ad23c13" (UID: "aa5b4806-95dc-4496-81cd-d66b3ad23c13"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.550955 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d") pod "aa5b4806-95dc-4496-81cd-d66b3ad23c13" (UID: "aa5b4806-95dc-4496-81cd-d66b3ad23c13"). InnerVolumeSpecName "web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.553784 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa5b4806-95dc-4496-81cd-d66b3ad23c13-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "aa5b4806-95dc-4496-81cd-d66b3ad23c13" (UID: "aa5b4806-95dc-4496-81cd-d66b3ad23c13"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.555721 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa5b4806-95dc-4496-81cd-d66b3ad23c13-config-out" (OuterVolumeSpecName: "config-out") pod "aa5b4806-95dc-4496-81cd-d66b3ad23c13" (UID: "aa5b4806-95dc-4496-81cd-d66b3ad23c13"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.555758 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-secret-combined-ca-bundle" (OuterVolumeSpecName: "secret-combined-ca-bundle") pod "aa5b4806-95dc-4496-81cd-d66b3ad23c13" (UID: "aa5b4806-95dc-4496-81cd-d66b3ad23c13"). InnerVolumeSpecName "secret-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.555721 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d" (OuterVolumeSpecName: "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d") pod "aa5b4806-95dc-4496-81cd-d66b3ad23c13" (UID: "aa5b4806-95dc-4496-81cd-d66b3ad23c13"). InnerVolumeSpecName "web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.555795 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "aa5b4806-95dc-4496-81cd-d66b3ad23c13" (UID: "aa5b4806-95dc-4496-81cd-d66b3ad23c13"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.555921 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa5b4806-95dc-4496-81cd-d66b3ad23c13-kube-api-access-r5ngs" (OuterVolumeSpecName: "kube-api-access-r5ngs") pod "aa5b4806-95dc-4496-81cd-d66b3ad23c13" (UID: "aa5b4806-95dc-4496-81cd-d66b3ad23c13"). InnerVolumeSpecName "kube-api-access-r5ngs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.556284 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-config" (OuterVolumeSpecName: "config") pod "aa5b4806-95dc-4496-81cd-d66b3ad23c13" (UID: "aa5b4806-95dc-4496-81cd-d66b3ad23c13"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.582577 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "aa5b4806-95dc-4496-81cd-d66b3ad23c13" (UID: "aa5b4806-95dc-4496-81cd-d66b3ad23c13"). InnerVolumeSpecName "pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.616041 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:30:59 crc kubenswrapper[4884]: E1202 02:30:59.616394 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.625285 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config" (OuterVolumeSpecName: "web-config") pod "aa5b4806-95dc-4496-81cd-d66b3ad23c13" (UID: "aa5b4806-95dc-4496-81cd-d66b3ad23c13"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.646795 4884 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/aa5b4806-95dc-4496-81cd-d66b3ad23c13-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.646833 4884 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.646845 4884 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.646857 4884 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.646870 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5ngs\" (UniqueName: \"kubernetes.io/projected/aa5b4806-95dc-4496-81cd-d66b3ad23c13-kube-api-access-r5ngs\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.646889 4884 reconciler_common.go:293] "Volume detached for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.646900 4884 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/aa5b4806-95dc-4496-81cd-d66b3ad23c13-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.646908 4884 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/aa5b4806-95dc-4496-81cd-d66b3ad23c13-config-out\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.646915 4884 reconciler_common.go:293] "Volume detached for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-secret-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.646924 4884 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/aa5b4806-95dc-4496-81cd-d66b3ad23c13-config\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.646953 4884 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") on node \"crc\" " Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.675476 4884 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.676566 4884 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05") on node "crc" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.748897 4884 reconciler_common.go:293] "Volume detached for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") on node \"crc\" DevicePath \"\"" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.968233 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"aa5b4806-95dc-4496-81cd-d66b3ad23c13","Type":"ContainerDied","Data":"7f609c966f2c9327164b9143efaf0eec277af4bb3a8fd991185c3f91876b8610"} Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.968292 4884 scope.go:117] "RemoveContainer" containerID="6103d609ead7e74a2f2a809ab023c79ca78a2835c48a603c317ecbf6f95d38bd" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.968310 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.992539 4884 scope.go:117] "RemoveContainer" containerID="22cae3ab5f411ea5f30ea9e8ad3ff278d413825ae977a8ee143fd766e1cf651a" Dec 02 02:30:59 crc kubenswrapper[4884]: I1202 02:30:59.998488 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.008153 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.026286 4884 scope.go:117] "RemoveContainer" containerID="ff2a72f07e42090c0d6e3254cf3b17dd9ef3b8f1b498858eb94632c5a0884931" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.038450 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 02 02:31:00 crc kubenswrapper[4884]: E1202 02:31:00.038955 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerName="thanos-sidecar" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.038977 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerName="thanos-sidecar" Dec 02 02:31:00 crc kubenswrapper[4884]: E1202 02:31:00.038995 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5" containerName="collect-profiles" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.039004 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5" containerName="collect-profiles" Dec 02 02:31:00 crc kubenswrapper[4884]: E1202 02:31:00.039028 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerName="config-reloader" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.039038 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerName="config-reloader" Dec 02 02:31:00 crc kubenswrapper[4884]: E1202 02:31:00.039068 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7098a534-9f0a-457d-a3d2-09249bd490ee" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.039077 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="7098a534-9f0a-457d-a3d2-09249bd490ee" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 02 02:31:00 crc kubenswrapper[4884]: E1202 02:31:00.039095 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerName="prometheus" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.039102 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerName="prometheus" Dec 02 02:31:00 crc kubenswrapper[4884]: E1202 02:31:00.039123 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerName="init-config-reloader" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.039132 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerName="init-config-reloader" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.039368 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5" containerName="collect-profiles" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.039544 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerName="thanos-sidecar" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.039564 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="7098a534-9f0a-457d-a3d2-09249bd490ee" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.039614 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerName="prometheus" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.039635 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" containerName="config-reloader" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.043816 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.047991 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-mcslp" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.048400 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.048824 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.049679 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.049867 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.059190 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.064255 4884 scope.go:117] "RemoveContainer" containerID="5b5f08beb105d5a87ecc726ef6fdf90d4b742b063251973973341d0df9951e9f" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.066546 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.155594 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.155641 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/827e7eb5-0a90-4677-9d1f-50185bf48b52-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.155692 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.155721 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-config\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.155831 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/827e7eb5-0a90-4677-9d1f-50185bf48b52-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.155857 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.155924 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.155958 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.155992 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/827e7eb5-0a90-4677-9d1f-50185bf48b52-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.156155 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.156219 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4r87\" (UniqueName: \"kubernetes.io/projected/827e7eb5-0a90-4677-9d1f-50185bf48b52-kube-api-access-b4r87\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.257878 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/827e7eb5-0a90-4677-9d1f-50185bf48b52-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.257928 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.257997 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.258023 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.258051 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/827e7eb5-0a90-4677-9d1f-50185bf48b52-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.258078 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.258097 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4r87\" (UniqueName: \"kubernetes.io/projected/827e7eb5-0a90-4677-9d1f-50185bf48b52-kube-api-access-b4r87\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.258148 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.258193 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/827e7eb5-0a90-4677-9d1f-50185bf48b52-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.258238 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.258263 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-config\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.259054 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/827e7eb5-0a90-4677-9d1f-50185bf48b52-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.263488 4884 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.263809 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1350ca4960854133c6cd5b98efbfa142de46bc322b82e157adbca982350a030b/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.263562 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.263573 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.264365 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.265008 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.265145 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/827e7eb5-0a90-4677-9d1f-50185bf48b52-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.265733 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.266278 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/827e7eb5-0a90-4677-9d1f-50185bf48b52-config\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.271995 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/827e7eb5-0a90-4677-9d1f-50185bf48b52-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.277071 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4r87\" (UniqueName: \"kubernetes.io/projected/827e7eb5-0a90-4677-9d1f-50185bf48b52-kube-api-access-b4r87\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.326939 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b494f58e-456c-4aa2-9313-1fba6b37ad05\") pod \"prometheus-metric-storage-0\" (UID: \"827e7eb5-0a90-4677-9d1f-50185bf48b52\") " pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.444485 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:00 crc kubenswrapper[4884]: I1202 02:31:00.974227 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 02 02:31:00 crc kubenswrapper[4884]: W1202 02:31:00.976046 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod827e7eb5_0a90_4677_9d1f_50185bf48b52.slice/crio-21ea0ba564a958c95b963f73921733e6d772940932c786bc321e977f78d2704a WatchSource:0}: Error finding container 21ea0ba564a958c95b963f73921733e6d772940932c786bc321e977f78d2704a: Status 404 returned error can't find the container with id 21ea0ba564a958c95b963f73921733e6d772940932c786bc321e977f78d2704a Dec 02 02:31:01 crc kubenswrapper[4884]: I1202 02:31:01.634889 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa5b4806-95dc-4496-81cd-d66b3ad23c13" path="/var/lib/kubelet/pods/aa5b4806-95dc-4496-81cd-d66b3ad23c13/volumes" Dec 02 02:31:01 crc kubenswrapper[4884]: I1202 02:31:01.996005 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"827e7eb5-0a90-4677-9d1f-50185bf48b52","Type":"ContainerStarted","Data":"21ea0ba564a958c95b963f73921733e6d772940932c786bc321e977f78d2704a"} Dec 02 02:31:06 crc kubenswrapper[4884]: I1202 02:31:06.046943 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"827e7eb5-0a90-4677-9d1f-50185bf48b52","Type":"ContainerStarted","Data":"9dbcdd703820d77db82856073d4c9487f22dcfd0afc0806050ff066ef691b9ee"} Dec 02 02:31:11 crc kubenswrapper[4884]: I1202 02:31:11.614949 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:31:11 crc kubenswrapper[4884]: E1202 02:31:11.615813 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:31:17 crc kubenswrapper[4884]: I1202 02:31:17.178241 4884 generic.go:334] "Generic (PLEG): container finished" podID="827e7eb5-0a90-4677-9d1f-50185bf48b52" containerID="9dbcdd703820d77db82856073d4c9487f22dcfd0afc0806050ff066ef691b9ee" exitCode=0 Dec 02 02:31:17 crc kubenswrapper[4884]: I1202 02:31:17.178357 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"827e7eb5-0a90-4677-9d1f-50185bf48b52","Type":"ContainerDied","Data":"9dbcdd703820d77db82856073d4c9487f22dcfd0afc0806050ff066ef691b9ee"} Dec 02 02:31:18 crc kubenswrapper[4884]: I1202 02:31:18.193316 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"827e7eb5-0a90-4677-9d1f-50185bf48b52","Type":"ContainerStarted","Data":"55207b4d16b428b485fdc406363a01dd87d7d0cdbcb9b85cbb1bcedc2251e482"} Dec 02 02:31:21 crc kubenswrapper[4884]: I1202 02:31:21.241537 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"827e7eb5-0a90-4677-9d1f-50185bf48b52","Type":"ContainerStarted","Data":"2cea999383c7f8077b553da5896e491dac63c41a945315f8b34052f53e5deece"} Dec 02 02:31:21 crc kubenswrapper[4884]: I1202 02:31:21.242073 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"827e7eb5-0a90-4677-9d1f-50185bf48b52","Type":"ContainerStarted","Data":"090bedc296f8c8c65763e53230458b4a8b397838b7d2ff3b15dbdb1ef1a99612"} Dec 02 02:31:21 crc kubenswrapper[4884]: I1202 02:31:21.303182 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=21.303151897 podStartE2EDuration="21.303151897s" podCreationTimestamp="2025-12-02 02:31:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 02:31:21.286710545 +0000 UTC m=+3297.962547489" watchObservedRunningTime="2025-12-02 02:31:21.303151897 +0000 UTC m=+3297.978988821" Dec 02 02:31:22 crc kubenswrapper[4884]: I1202 02:31:22.614930 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:31:22 crc kubenswrapper[4884]: E1202 02:31:22.615517 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:31:25 crc kubenswrapper[4884]: I1202 02:31:25.445764 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:30 crc kubenswrapper[4884]: I1202 02:31:30.444700 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:30 crc kubenswrapper[4884]: I1202 02:31:30.456661 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:31 crc kubenswrapper[4884]: I1202 02:31:31.362507 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 02 02:31:37 crc kubenswrapper[4884]: I1202 02:31:37.616715 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:31:37 crc kubenswrapper[4884]: E1202 02:31:37.617615 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:31:49 crc kubenswrapper[4884]: I1202 02:31:49.614558 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:31:49 crc kubenswrapper[4884]: E1202 02:31:49.615524 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.655737 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.658162 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.678994 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.693154 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-v2jd6" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.693378 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.693541 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.693641 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.743587 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56afebdd-31ad-4882-b405-53ada83cbaea-config-data\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.743734 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.743812 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.743999 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/56afebdd-31ad-4882-b405-53ada83cbaea-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.744618 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpbvf\" (UniqueName: \"kubernetes.io/projected/56afebdd-31ad-4882-b405-53ada83cbaea-kube-api-access-bpbvf\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.744795 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/56afebdd-31ad-4882-b405-53ada83cbaea-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.744839 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.744973 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.745024 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/56afebdd-31ad-4882-b405-53ada83cbaea-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.846805 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.846875 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/56afebdd-31ad-4882-b405-53ada83cbaea-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.846920 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56afebdd-31ad-4882-b405-53ada83cbaea-config-data\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.846965 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.847000 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.847032 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/56afebdd-31ad-4882-b405-53ada83cbaea-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.847081 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpbvf\" (UniqueName: \"kubernetes.io/projected/56afebdd-31ad-4882-b405-53ada83cbaea-kube-api-access-bpbvf\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.847357 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/56afebdd-31ad-4882-b405-53ada83cbaea-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.847399 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.849031 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/56afebdd-31ad-4882-b405-53ada83cbaea-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.849880 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/56afebdd-31ad-4882-b405-53ada83cbaea-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.850104 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/56afebdd-31ad-4882-b405-53ada83cbaea-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.850563 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.856717 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.857225 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.857942 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.858446 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56afebdd-31ad-4882-b405-53ada83cbaea-config-data\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.867507 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpbvf\" (UniqueName: \"kubernetes.io/projected/56afebdd-31ad-4882-b405-53ada83cbaea-kube-api-access-bpbvf\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:50 crc kubenswrapper[4884]: I1202 02:31:50.884041 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " pod="openstack/tempest-tests-tempest" Dec 02 02:31:51 crc kubenswrapper[4884]: I1202 02:31:51.018170 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 02 02:31:51 crc kubenswrapper[4884]: I1202 02:31:51.687691 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 02 02:31:51 crc kubenswrapper[4884]: W1202 02:31:51.695042 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56afebdd_31ad_4882_b405_53ada83cbaea.slice/crio-ea1f47aea8b46bc72b37b2869347d735dbe1e7654cc60ebd7edaf36f648094ba WatchSource:0}: Error finding container ea1f47aea8b46bc72b37b2869347d735dbe1e7654cc60ebd7edaf36f648094ba: Status 404 returned error can't find the container with id ea1f47aea8b46bc72b37b2869347d735dbe1e7654cc60ebd7edaf36f648094ba Dec 02 02:31:52 crc kubenswrapper[4884]: I1202 02:31:52.583978 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"56afebdd-31ad-4882-b405-53ada83cbaea","Type":"ContainerStarted","Data":"ea1f47aea8b46bc72b37b2869347d735dbe1e7654cc60ebd7edaf36f648094ba"} Dec 02 02:32:03 crc kubenswrapper[4884]: I1202 02:32:03.626094 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:32:03 crc kubenswrapper[4884]: E1202 02:32:03.626811 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:32:03 crc kubenswrapper[4884]: I1202 02:32:03.695671 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"56afebdd-31ad-4882-b405-53ada83cbaea","Type":"ContainerStarted","Data":"9ed0eab2809a6d980eb0bd4cd922380106ecffb5c9cbaffa831c89315a5b8a9d"} Dec 02 02:32:03 crc kubenswrapper[4884]: I1202 02:32:03.729410 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.973973414 podStartE2EDuration="14.729387274s" podCreationTimestamp="2025-12-02 02:31:49 +0000 UTC" firstStartedPulling="2025-12-02 02:31:51.698126006 +0000 UTC m=+3328.373962920" lastFinishedPulling="2025-12-02 02:32:02.453539896 +0000 UTC m=+3339.129376780" observedRunningTime="2025-12-02 02:32:03.723530641 +0000 UTC m=+3340.399367535" watchObservedRunningTime="2025-12-02 02:32:03.729387274 +0000 UTC m=+3340.405224198" Dec 02 02:32:17 crc kubenswrapper[4884]: I1202 02:32:17.614502 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:32:17 crc kubenswrapper[4884]: E1202 02:32:17.615279 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:32:31 crc kubenswrapper[4884]: I1202 02:32:31.615036 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:32:31 crc kubenswrapper[4884]: E1202 02:32:31.616461 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:32:45 crc kubenswrapper[4884]: I1202 02:32:45.614881 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:32:45 crc kubenswrapper[4884]: E1202 02:32:45.616024 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:32:56 crc kubenswrapper[4884]: I1202 02:32:56.614878 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:32:56 crc kubenswrapper[4884]: E1202 02:32:56.616181 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:33:08 crc kubenswrapper[4884]: I1202 02:33:08.614427 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:33:08 crc kubenswrapper[4884]: E1202 02:33:08.615261 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:33:23 crc kubenswrapper[4884]: I1202 02:33:23.632060 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:33:24 crc kubenswrapper[4884]: I1202 02:33:24.740231 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"1171eeb0bae3a98055f356e1079ce475be63ba1276ce35ce9e5b5ec76d526801"} Dec 02 02:34:24 crc kubenswrapper[4884]: I1202 02:34:24.621861 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n9v7h"] Dec 02 02:34:24 crc kubenswrapper[4884]: I1202 02:34:24.627574 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:24 crc kubenswrapper[4884]: I1202 02:34:24.645946 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n9v7h"] Dec 02 02:34:24 crc kubenswrapper[4884]: I1202 02:34:24.688628 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fb79a85-a86f-4551-ae60-7822752e7f6c-utilities\") pod \"community-operators-n9v7h\" (UID: \"1fb79a85-a86f-4551-ae60-7822752e7f6c\") " pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:24 crc kubenswrapper[4884]: I1202 02:34:24.688917 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fb79a85-a86f-4551-ae60-7822752e7f6c-catalog-content\") pod \"community-operators-n9v7h\" (UID: \"1fb79a85-a86f-4551-ae60-7822752e7f6c\") " pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:24 crc kubenswrapper[4884]: I1202 02:34:24.689179 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jktnw\" (UniqueName: \"kubernetes.io/projected/1fb79a85-a86f-4551-ae60-7822752e7f6c-kube-api-access-jktnw\") pod \"community-operators-n9v7h\" (UID: \"1fb79a85-a86f-4551-ae60-7822752e7f6c\") " pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:24 crc kubenswrapper[4884]: I1202 02:34:24.790159 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fb79a85-a86f-4551-ae60-7822752e7f6c-catalog-content\") pod \"community-operators-n9v7h\" (UID: \"1fb79a85-a86f-4551-ae60-7822752e7f6c\") " pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:24 crc kubenswrapper[4884]: I1202 02:34:24.790230 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jktnw\" (UniqueName: \"kubernetes.io/projected/1fb79a85-a86f-4551-ae60-7822752e7f6c-kube-api-access-jktnw\") pod \"community-operators-n9v7h\" (UID: \"1fb79a85-a86f-4551-ae60-7822752e7f6c\") " pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:24 crc kubenswrapper[4884]: I1202 02:34:24.790281 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fb79a85-a86f-4551-ae60-7822752e7f6c-utilities\") pod \"community-operators-n9v7h\" (UID: \"1fb79a85-a86f-4551-ae60-7822752e7f6c\") " pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:24 crc kubenswrapper[4884]: I1202 02:34:24.790770 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fb79a85-a86f-4551-ae60-7822752e7f6c-utilities\") pod \"community-operators-n9v7h\" (UID: \"1fb79a85-a86f-4551-ae60-7822752e7f6c\") " pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:24 crc kubenswrapper[4884]: I1202 02:34:24.790981 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fb79a85-a86f-4551-ae60-7822752e7f6c-catalog-content\") pod \"community-operators-n9v7h\" (UID: \"1fb79a85-a86f-4551-ae60-7822752e7f6c\") " pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:24 crc kubenswrapper[4884]: I1202 02:34:24.809207 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jktnw\" (UniqueName: \"kubernetes.io/projected/1fb79a85-a86f-4551-ae60-7822752e7f6c-kube-api-access-jktnw\") pod \"community-operators-n9v7h\" (UID: \"1fb79a85-a86f-4551-ae60-7822752e7f6c\") " pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:24 crc kubenswrapper[4884]: I1202 02:34:24.982530 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:25 crc kubenswrapper[4884]: W1202 02:34:25.563918 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fb79a85_a86f_4551_ae60_7822752e7f6c.slice/crio-ca9e489e8b688b208fa93ab6d644912ecb716ededdb84566e3e425ed19df5e44 WatchSource:0}: Error finding container ca9e489e8b688b208fa93ab6d644912ecb716ededdb84566e3e425ed19df5e44: Status 404 returned error can't find the container with id ca9e489e8b688b208fa93ab6d644912ecb716ededdb84566e3e425ed19df5e44 Dec 02 02:34:25 crc kubenswrapper[4884]: I1202 02:34:25.565665 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n9v7h"] Dec 02 02:34:26 crc kubenswrapper[4884]: I1202 02:34:26.512175 4884 generic.go:334] "Generic (PLEG): container finished" podID="1fb79a85-a86f-4551-ae60-7822752e7f6c" containerID="8cc73a3043e631a1c529c7c369622a26a49adaa6612a32e6dc442b09e4f4b355" exitCode=0 Dec 02 02:34:26 crc kubenswrapper[4884]: I1202 02:34:26.512386 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9v7h" event={"ID":"1fb79a85-a86f-4551-ae60-7822752e7f6c","Type":"ContainerDied","Data":"8cc73a3043e631a1c529c7c369622a26a49adaa6612a32e6dc442b09e4f4b355"} Dec 02 02:34:26 crc kubenswrapper[4884]: I1202 02:34:26.512639 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9v7h" event={"ID":"1fb79a85-a86f-4551-ae60-7822752e7f6c","Type":"ContainerStarted","Data":"ca9e489e8b688b208fa93ab6d644912ecb716ededdb84566e3e425ed19df5e44"} Dec 02 02:34:26 crc kubenswrapper[4884]: I1202 02:34:26.520213 4884 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 02:34:27 crc kubenswrapper[4884]: I1202 02:34:27.525847 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9v7h" event={"ID":"1fb79a85-a86f-4551-ae60-7822752e7f6c","Type":"ContainerStarted","Data":"1fc0ef83e59da7692a1d77ab5bc408c65e56dade6d15e1d836b4da76a60c88ee"} Dec 02 02:34:28 crc kubenswrapper[4884]: I1202 02:34:28.537055 4884 generic.go:334] "Generic (PLEG): container finished" podID="1fb79a85-a86f-4551-ae60-7822752e7f6c" containerID="1fc0ef83e59da7692a1d77ab5bc408c65e56dade6d15e1d836b4da76a60c88ee" exitCode=0 Dec 02 02:34:28 crc kubenswrapper[4884]: I1202 02:34:28.537140 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9v7h" event={"ID":"1fb79a85-a86f-4551-ae60-7822752e7f6c","Type":"ContainerDied","Data":"1fc0ef83e59da7692a1d77ab5bc408c65e56dade6d15e1d836b4da76a60c88ee"} Dec 02 02:34:29 crc kubenswrapper[4884]: I1202 02:34:29.566585 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9v7h" event={"ID":"1fb79a85-a86f-4551-ae60-7822752e7f6c","Type":"ContainerStarted","Data":"086a95290cbe2c534c93379a64df40abb72f507b514dd66f6b7854b800ff9339"} Dec 02 02:34:29 crc kubenswrapper[4884]: I1202 02:34:29.610375 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n9v7h" podStartSLOduration=3.050644098 podStartE2EDuration="5.610341041s" podCreationTimestamp="2025-12-02 02:34:24 +0000 UTC" firstStartedPulling="2025-12-02 02:34:26.519974985 +0000 UTC m=+3483.195811859" lastFinishedPulling="2025-12-02 02:34:29.079671918 +0000 UTC m=+3485.755508802" observedRunningTime="2025-12-02 02:34:29.594529414 +0000 UTC m=+3486.270366358" watchObservedRunningTime="2025-12-02 02:34:29.610341041 +0000 UTC m=+3486.286177965" Dec 02 02:34:34 crc kubenswrapper[4884]: I1202 02:34:34.983361 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:34 crc kubenswrapper[4884]: I1202 02:34:34.984593 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:35 crc kubenswrapper[4884]: I1202 02:34:35.062132 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:35 crc kubenswrapper[4884]: I1202 02:34:35.698981 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:35 crc kubenswrapper[4884]: I1202 02:34:35.749493 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n9v7h"] Dec 02 02:34:37 crc kubenswrapper[4884]: I1202 02:34:37.660861 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n9v7h" podUID="1fb79a85-a86f-4551-ae60-7822752e7f6c" containerName="registry-server" containerID="cri-o://086a95290cbe2c534c93379a64df40abb72f507b514dd66f6b7854b800ff9339" gracePeriod=2 Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.168663 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.271762 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jktnw\" (UniqueName: \"kubernetes.io/projected/1fb79a85-a86f-4551-ae60-7822752e7f6c-kube-api-access-jktnw\") pod \"1fb79a85-a86f-4551-ae60-7822752e7f6c\" (UID: \"1fb79a85-a86f-4551-ae60-7822752e7f6c\") " Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.272032 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fb79a85-a86f-4551-ae60-7822752e7f6c-utilities\") pod \"1fb79a85-a86f-4551-ae60-7822752e7f6c\" (UID: \"1fb79a85-a86f-4551-ae60-7822752e7f6c\") " Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.272077 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fb79a85-a86f-4551-ae60-7822752e7f6c-catalog-content\") pod \"1fb79a85-a86f-4551-ae60-7822752e7f6c\" (UID: \"1fb79a85-a86f-4551-ae60-7822752e7f6c\") " Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.272939 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fb79a85-a86f-4551-ae60-7822752e7f6c-utilities" (OuterVolumeSpecName: "utilities") pod "1fb79a85-a86f-4551-ae60-7822752e7f6c" (UID: "1fb79a85-a86f-4551-ae60-7822752e7f6c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.276967 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fb79a85-a86f-4551-ae60-7822752e7f6c-kube-api-access-jktnw" (OuterVolumeSpecName: "kube-api-access-jktnw") pod "1fb79a85-a86f-4551-ae60-7822752e7f6c" (UID: "1fb79a85-a86f-4551-ae60-7822752e7f6c"). InnerVolumeSpecName "kube-api-access-jktnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.320443 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fb79a85-a86f-4551-ae60-7822752e7f6c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1fb79a85-a86f-4551-ae60-7822752e7f6c" (UID: "1fb79a85-a86f-4551-ae60-7822752e7f6c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.376211 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jktnw\" (UniqueName: \"kubernetes.io/projected/1fb79a85-a86f-4551-ae60-7822752e7f6c-kube-api-access-jktnw\") on node \"crc\" DevicePath \"\"" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.376277 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1fb79a85-a86f-4551-ae60-7822752e7f6c-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.376309 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1fb79a85-a86f-4551-ae60-7822752e7f6c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.673522 4884 generic.go:334] "Generic (PLEG): container finished" podID="1fb79a85-a86f-4551-ae60-7822752e7f6c" containerID="086a95290cbe2c534c93379a64df40abb72f507b514dd66f6b7854b800ff9339" exitCode=0 Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.673584 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n9v7h" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.673604 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9v7h" event={"ID":"1fb79a85-a86f-4551-ae60-7822752e7f6c","Type":"ContainerDied","Data":"086a95290cbe2c534c93379a64df40abb72f507b514dd66f6b7854b800ff9339"} Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.676815 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n9v7h" event={"ID":"1fb79a85-a86f-4551-ae60-7822752e7f6c","Type":"ContainerDied","Data":"ca9e489e8b688b208fa93ab6d644912ecb716ededdb84566e3e425ed19df5e44"} Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.676852 4884 scope.go:117] "RemoveContainer" containerID="086a95290cbe2c534c93379a64df40abb72f507b514dd66f6b7854b800ff9339" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.712690 4884 scope.go:117] "RemoveContainer" containerID="1fc0ef83e59da7692a1d77ab5bc408c65e56dade6d15e1d836b4da76a60c88ee" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.718927 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n9v7h"] Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.735146 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n9v7h"] Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.748530 4884 scope.go:117] "RemoveContainer" containerID="8cc73a3043e631a1c529c7c369622a26a49adaa6612a32e6dc442b09e4f4b355" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.807604 4884 scope.go:117] "RemoveContainer" containerID="086a95290cbe2c534c93379a64df40abb72f507b514dd66f6b7854b800ff9339" Dec 02 02:34:38 crc kubenswrapper[4884]: E1202 02:34:38.808128 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"086a95290cbe2c534c93379a64df40abb72f507b514dd66f6b7854b800ff9339\": container with ID starting with 086a95290cbe2c534c93379a64df40abb72f507b514dd66f6b7854b800ff9339 not found: ID does not exist" containerID="086a95290cbe2c534c93379a64df40abb72f507b514dd66f6b7854b800ff9339" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.808159 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"086a95290cbe2c534c93379a64df40abb72f507b514dd66f6b7854b800ff9339"} err="failed to get container status \"086a95290cbe2c534c93379a64df40abb72f507b514dd66f6b7854b800ff9339\": rpc error: code = NotFound desc = could not find container \"086a95290cbe2c534c93379a64df40abb72f507b514dd66f6b7854b800ff9339\": container with ID starting with 086a95290cbe2c534c93379a64df40abb72f507b514dd66f6b7854b800ff9339 not found: ID does not exist" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.808183 4884 scope.go:117] "RemoveContainer" containerID="1fc0ef83e59da7692a1d77ab5bc408c65e56dade6d15e1d836b4da76a60c88ee" Dec 02 02:34:38 crc kubenswrapper[4884]: E1202 02:34:38.809283 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fc0ef83e59da7692a1d77ab5bc408c65e56dade6d15e1d836b4da76a60c88ee\": container with ID starting with 1fc0ef83e59da7692a1d77ab5bc408c65e56dade6d15e1d836b4da76a60c88ee not found: ID does not exist" containerID="1fc0ef83e59da7692a1d77ab5bc408c65e56dade6d15e1d836b4da76a60c88ee" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.809307 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fc0ef83e59da7692a1d77ab5bc408c65e56dade6d15e1d836b4da76a60c88ee"} err="failed to get container status \"1fc0ef83e59da7692a1d77ab5bc408c65e56dade6d15e1d836b4da76a60c88ee\": rpc error: code = NotFound desc = could not find container \"1fc0ef83e59da7692a1d77ab5bc408c65e56dade6d15e1d836b4da76a60c88ee\": container with ID starting with 1fc0ef83e59da7692a1d77ab5bc408c65e56dade6d15e1d836b4da76a60c88ee not found: ID does not exist" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.809325 4884 scope.go:117] "RemoveContainer" containerID="8cc73a3043e631a1c529c7c369622a26a49adaa6612a32e6dc442b09e4f4b355" Dec 02 02:34:38 crc kubenswrapper[4884]: E1202 02:34:38.809825 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cc73a3043e631a1c529c7c369622a26a49adaa6612a32e6dc442b09e4f4b355\": container with ID starting with 8cc73a3043e631a1c529c7c369622a26a49adaa6612a32e6dc442b09e4f4b355 not found: ID does not exist" containerID="8cc73a3043e631a1c529c7c369622a26a49adaa6612a32e6dc442b09e4f4b355" Dec 02 02:34:38 crc kubenswrapper[4884]: I1202 02:34:38.809871 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cc73a3043e631a1c529c7c369622a26a49adaa6612a32e6dc442b09e4f4b355"} err="failed to get container status \"8cc73a3043e631a1c529c7c369622a26a49adaa6612a32e6dc442b09e4f4b355\": rpc error: code = NotFound desc = could not find container \"8cc73a3043e631a1c529c7c369622a26a49adaa6612a32e6dc442b09e4f4b355\": container with ID starting with 8cc73a3043e631a1c529c7c369622a26a49adaa6612a32e6dc442b09e4f4b355 not found: ID does not exist" Dec 02 02:34:39 crc kubenswrapper[4884]: I1202 02:34:39.629130 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fb79a85-a86f-4551-ae60-7822752e7f6c" path="/var/lib/kubelet/pods/1fb79a85-a86f-4551-ae60-7822752e7f6c/volumes" Dec 02 02:34:57 crc kubenswrapper[4884]: I1202 02:34:57.828060 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dtjrn"] Dec 02 02:34:57 crc kubenswrapper[4884]: E1202 02:34:57.829218 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb79a85-a86f-4551-ae60-7822752e7f6c" containerName="registry-server" Dec 02 02:34:57 crc kubenswrapper[4884]: I1202 02:34:57.829240 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb79a85-a86f-4551-ae60-7822752e7f6c" containerName="registry-server" Dec 02 02:34:57 crc kubenswrapper[4884]: E1202 02:34:57.829255 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb79a85-a86f-4551-ae60-7822752e7f6c" containerName="extract-content" Dec 02 02:34:57 crc kubenswrapper[4884]: I1202 02:34:57.829265 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb79a85-a86f-4551-ae60-7822752e7f6c" containerName="extract-content" Dec 02 02:34:57 crc kubenswrapper[4884]: E1202 02:34:57.829331 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb79a85-a86f-4551-ae60-7822752e7f6c" containerName="extract-utilities" Dec 02 02:34:57 crc kubenswrapper[4884]: I1202 02:34:57.829344 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb79a85-a86f-4551-ae60-7822752e7f6c" containerName="extract-utilities" Dec 02 02:34:57 crc kubenswrapper[4884]: I1202 02:34:57.829636 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fb79a85-a86f-4551-ae60-7822752e7f6c" containerName="registry-server" Dec 02 02:34:57 crc kubenswrapper[4884]: I1202 02:34:57.833185 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:34:57 crc kubenswrapper[4884]: I1202 02:34:57.851915 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtjrn"] Dec 02 02:34:58 crc kubenswrapper[4884]: I1202 02:34:58.008038 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb115e59-faa3-454c-958c-cbeac3d80750-catalog-content\") pod \"redhat-marketplace-dtjrn\" (UID: \"bb115e59-faa3-454c-958c-cbeac3d80750\") " pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:34:58 crc kubenswrapper[4884]: I1202 02:34:58.008089 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb115e59-faa3-454c-958c-cbeac3d80750-utilities\") pod \"redhat-marketplace-dtjrn\" (UID: \"bb115e59-faa3-454c-958c-cbeac3d80750\") " pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:34:58 crc kubenswrapper[4884]: I1202 02:34:58.008207 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z7n7\" (UniqueName: \"kubernetes.io/projected/bb115e59-faa3-454c-958c-cbeac3d80750-kube-api-access-8z7n7\") pod \"redhat-marketplace-dtjrn\" (UID: \"bb115e59-faa3-454c-958c-cbeac3d80750\") " pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:34:58 crc kubenswrapper[4884]: I1202 02:34:58.109809 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z7n7\" (UniqueName: \"kubernetes.io/projected/bb115e59-faa3-454c-958c-cbeac3d80750-kube-api-access-8z7n7\") pod \"redhat-marketplace-dtjrn\" (UID: \"bb115e59-faa3-454c-958c-cbeac3d80750\") " pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:34:58 crc kubenswrapper[4884]: I1202 02:34:58.109915 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb115e59-faa3-454c-958c-cbeac3d80750-catalog-content\") pod \"redhat-marketplace-dtjrn\" (UID: \"bb115e59-faa3-454c-958c-cbeac3d80750\") " pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:34:58 crc kubenswrapper[4884]: I1202 02:34:58.109941 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb115e59-faa3-454c-958c-cbeac3d80750-utilities\") pod \"redhat-marketplace-dtjrn\" (UID: \"bb115e59-faa3-454c-958c-cbeac3d80750\") " pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:34:58 crc kubenswrapper[4884]: I1202 02:34:58.110438 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb115e59-faa3-454c-958c-cbeac3d80750-utilities\") pod \"redhat-marketplace-dtjrn\" (UID: \"bb115e59-faa3-454c-958c-cbeac3d80750\") " pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:34:58 crc kubenswrapper[4884]: I1202 02:34:58.110644 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb115e59-faa3-454c-958c-cbeac3d80750-catalog-content\") pod \"redhat-marketplace-dtjrn\" (UID: \"bb115e59-faa3-454c-958c-cbeac3d80750\") " pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:34:58 crc kubenswrapper[4884]: I1202 02:34:58.133999 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z7n7\" (UniqueName: \"kubernetes.io/projected/bb115e59-faa3-454c-958c-cbeac3d80750-kube-api-access-8z7n7\") pod \"redhat-marketplace-dtjrn\" (UID: \"bb115e59-faa3-454c-958c-cbeac3d80750\") " pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:34:58 crc kubenswrapper[4884]: I1202 02:34:58.186117 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:34:58 crc kubenswrapper[4884]: I1202 02:34:58.658415 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtjrn"] Dec 02 02:34:59 crc kubenswrapper[4884]: I1202 02:34:59.013443 4884 generic.go:334] "Generic (PLEG): container finished" podID="bb115e59-faa3-454c-958c-cbeac3d80750" containerID="d85b635aeb6f72e8242f477716fbb2c27140318b1a857e55ab194a1a21146c2e" exitCode=0 Dec 02 02:34:59 crc kubenswrapper[4884]: I1202 02:34:59.013574 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtjrn" event={"ID":"bb115e59-faa3-454c-958c-cbeac3d80750","Type":"ContainerDied","Data":"d85b635aeb6f72e8242f477716fbb2c27140318b1a857e55ab194a1a21146c2e"} Dec 02 02:34:59 crc kubenswrapper[4884]: I1202 02:34:59.013897 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtjrn" event={"ID":"bb115e59-faa3-454c-958c-cbeac3d80750","Type":"ContainerStarted","Data":"238eeed74cd2503dde482c299106fca973382d7159e8b5566062d4dccc3e8fb7"} Dec 02 02:35:00 crc kubenswrapper[4884]: I1202 02:35:00.024320 4884 generic.go:334] "Generic (PLEG): container finished" podID="bb115e59-faa3-454c-958c-cbeac3d80750" containerID="a4be986fdf87d54291b4943b846595a771cbe4cee6284d4a281831d6c34af916" exitCode=0 Dec 02 02:35:00 crc kubenswrapper[4884]: I1202 02:35:00.024432 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtjrn" event={"ID":"bb115e59-faa3-454c-958c-cbeac3d80750","Type":"ContainerDied","Data":"a4be986fdf87d54291b4943b846595a771cbe4cee6284d4a281831d6c34af916"} Dec 02 02:35:01 crc kubenswrapper[4884]: I1202 02:35:01.036013 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtjrn" event={"ID":"bb115e59-faa3-454c-958c-cbeac3d80750","Type":"ContainerStarted","Data":"131e79c6f61890a6162535e8374a8e6bb317e69b1ca656f49c5f9f1927211657"} Dec 02 02:35:01 crc kubenswrapper[4884]: I1202 02:35:01.071332 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dtjrn" podStartSLOduration=2.483489422 podStartE2EDuration="4.071307857s" podCreationTimestamp="2025-12-02 02:34:57 +0000 UTC" firstStartedPulling="2025-12-02 02:34:59.015356338 +0000 UTC m=+3515.691193222" lastFinishedPulling="2025-12-02 02:35:00.603174773 +0000 UTC m=+3517.279011657" observedRunningTime="2025-12-02 02:35:01.061887427 +0000 UTC m=+3517.737724351" watchObservedRunningTime="2025-12-02 02:35:01.071307857 +0000 UTC m=+3517.747144751" Dec 02 02:35:08 crc kubenswrapper[4884]: I1202 02:35:08.186495 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:35:08 crc kubenswrapper[4884]: I1202 02:35:08.187166 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:35:08 crc kubenswrapper[4884]: I1202 02:35:08.272944 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:35:09 crc kubenswrapper[4884]: I1202 02:35:09.174710 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:35:09 crc kubenswrapper[4884]: I1202 02:35:09.235692 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtjrn"] Dec 02 02:35:11 crc kubenswrapper[4884]: I1202 02:35:11.151240 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dtjrn" podUID="bb115e59-faa3-454c-958c-cbeac3d80750" containerName="registry-server" containerID="cri-o://131e79c6f61890a6162535e8374a8e6bb317e69b1ca656f49c5f9f1927211657" gracePeriod=2 Dec 02 02:35:11 crc kubenswrapper[4884]: I1202 02:35:11.707317 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:35:11 crc kubenswrapper[4884]: I1202 02:35:11.732023 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z7n7\" (UniqueName: \"kubernetes.io/projected/bb115e59-faa3-454c-958c-cbeac3d80750-kube-api-access-8z7n7\") pod \"bb115e59-faa3-454c-958c-cbeac3d80750\" (UID: \"bb115e59-faa3-454c-958c-cbeac3d80750\") " Dec 02 02:35:11 crc kubenswrapper[4884]: I1202 02:35:11.732105 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb115e59-faa3-454c-958c-cbeac3d80750-catalog-content\") pod \"bb115e59-faa3-454c-958c-cbeac3d80750\" (UID: \"bb115e59-faa3-454c-958c-cbeac3d80750\") " Dec 02 02:35:11 crc kubenswrapper[4884]: I1202 02:35:11.732302 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb115e59-faa3-454c-958c-cbeac3d80750-utilities\") pod \"bb115e59-faa3-454c-958c-cbeac3d80750\" (UID: \"bb115e59-faa3-454c-958c-cbeac3d80750\") " Dec 02 02:35:11 crc kubenswrapper[4884]: I1202 02:35:11.734183 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb115e59-faa3-454c-958c-cbeac3d80750-utilities" (OuterVolumeSpecName: "utilities") pod "bb115e59-faa3-454c-958c-cbeac3d80750" (UID: "bb115e59-faa3-454c-958c-cbeac3d80750"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:35:11 crc kubenswrapper[4884]: I1202 02:35:11.743843 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb115e59-faa3-454c-958c-cbeac3d80750-kube-api-access-8z7n7" (OuterVolumeSpecName: "kube-api-access-8z7n7") pod "bb115e59-faa3-454c-958c-cbeac3d80750" (UID: "bb115e59-faa3-454c-958c-cbeac3d80750"). InnerVolumeSpecName "kube-api-access-8z7n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:35:11 crc kubenswrapper[4884]: I1202 02:35:11.761947 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb115e59-faa3-454c-958c-cbeac3d80750-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb115e59-faa3-454c-958c-cbeac3d80750" (UID: "bb115e59-faa3-454c-958c-cbeac3d80750"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:35:11 crc kubenswrapper[4884]: I1202 02:35:11.835064 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z7n7\" (UniqueName: \"kubernetes.io/projected/bb115e59-faa3-454c-958c-cbeac3d80750-kube-api-access-8z7n7\") on node \"crc\" DevicePath \"\"" Dec 02 02:35:11 crc kubenswrapper[4884]: I1202 02:35:11.835318 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb115e59-faa3-454c-958c-cbeac3d80750-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:35:11 crc kubenswrapper[4884]: I1202 02:35:11.835439 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb115e59-faa3-454c-958c-cbeac3d80750-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:35:12 crc kubenswrapper[4884]: I1202 02:35:12.165562 4884 generic.go:334] "Generic (PLEG): container finished" podID="bb115e59-faa3-454c-958c-cbeac3d80750" containerID="131e79c6f61890a6162535e8374a8e6bb317e69b1ca656f49c5f9f1927211657" exitCode=0 Dec 02 02:35:12 crc kubenswrapper[4884]: I1202 02:35:12.165611 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtjrn" event={"ID":"bb115e59-faa3-454c-958c-cbeac3d80750","Type":"ContainerDied","Data":"131e79c6f61890a6162535e8374a8e6bb317e69b1ca656f49c5f9f1927211657"} Dec 02 02:35:12 crc kubenswrapper[4884]: I1202 02:35:12.165667 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dtjrn" event={"ID":"bb115e59-faa3-454c-958c-cbeac3d80750","Type":"ContainerDied","Data":"238eeed74cd2503dde482c299106fca973382d7159e8b5566062d4dccc3e8fb7"} Dec 02 02:35:12 crc kubenswrapper[4884]: I1202 02:35:12.165681 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dtjrn" Dec 02 02:35:12 crc kubenswrapper[4884]: I1202 02:35:12.165699 4884 scope.go:117] "RemoveContainer" containerID="131e79c6f61890a6162535e8374a8e6bb317e69b1ca656f49c5f9f1927211657" Dec 02 02:35:12 crc kubenswrapper[4884]: I1202 02:35:12.210035 4884 scope.go:117] "RemoveContainer" containerID="a4be986fdf87d54291b4943b846595a771cbe4cee6284d4a281831d6c34af916" Dec 02 02:35:12 crc kubenswrapper[4884]: I1202 02:35:12.251257 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtjrn"] Dec 02 02:35:12 crc kubenswrapper[4884]: I1202 02:35:12.259728 4884 scope.go:117] "RemoveContainer" containerID="d85b635aeb6f72e8242f477716fbb2c27140318b1a857e55ab194a1a21146c2e" Dec 02 02:35:12 crc kubenswrapper[4884]: I1202 02:35:12.268544 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dtjrn"] Dec 02 02:35:12 crc kubenswrapper[4884]: I1202 02:35:12.305135 4884 scope.go:117] "RemoveContainer" containerID="131e79c6f61890a6162535e8374a8e6bb317e69b1ca656f49c5f9f1927211657" Dec 02 02:35:12 crc kubenswrapper[4884]: E1202 02:35:12.307234 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"131e79c6f61890a6162535e8374a8e6bb317e69b1ca656f49c5f9f1927211657\": container with ID starting with 131e79c6f61890a6162535e8374a8e6bb317e69b1ca656f49c5f9f1927211657 not found: ID does not exist" containerID="131e79c6f61890a6162535e8374a8e6bb317e69b1ca656f49c5f9f1927211657" Dec 02 02:35:12 crc kubenswrapper[4884]: I1202 02:35:12.307284 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"131e79c6f61890a6162535e8374a8e6bb317e69b1ca656f49c5f9f1927211657"} err="failed to get container status \"131e79c6f61890a6162535e8374a8e6bb317e69b1ca656f49c5f9f1927211657\": rpc error: code = NotFound desc = could not find container \"131e79c6f61890a6162535e8374a8e6bb317e69b1ca656f49c5f9f1927211657\": container with ID starting with 131e79c6f61890a6162535e8374a8e6bb317e69b1ca656f49c5f9f1927211657 not found: ID does not exist" Dec 02 02:35:12 crc kubenswrapper[4884]: I1202 02:35:12.307319 4884 scope.go:117] "RemoveContainer" containerID="a4be986fdf87d54291b4943b846595a771cbe4cee6284d4a281831d6c34af916" Dec 02 02:35:12 crc kubenswrapper[4884]: E1202 02:35:12.307671 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4be986fdf87d54291b4943b846595a771cbe4cee6284d4a281831d6c34af916\": container with ID starting with a4be986fdf87d54291b4943b846595a771cbe4cee6284d4a281831d6c34af916 not found: ID does not exist" containerID="a4be986fdf87d54291b4943b846595a771cbe4cee6284d4a281831d6c34af916" Dec 02 02:35:12 crc kubenswrapper[4884]: I1202 02:35:12.307709 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4be986fdf87d54291b4943b846595a771cbe4cee6284d4a281831d6c34af916"} err="failed to get container status \"a4be986fdf87d54291b4943b846595a771cbe4cee6284d4a281831d6c34af916\": rpc error: code = NotFound desc = could not find container \"a4be986fdf87d54291b4943b846595a771cbe4cee6284d4a281831d6c34af916\": container with ID starting with a4be986fdf87d54291b4943b846595a771cbe4cee6284d4a281831d6c34af916 not found: ID does not exist" Dec 02 02:35:12 crc kubenswrapper[4884]: I1202 02:35:12.307735 4884 scope.go:117] "RemoveContainer" containerID="d85b635aeb6f72e8242f477716fbb2c27140318b1a857e55ab194a1a21146c2e" Dec 02 02:35:12 crc kubenswrapper[4884]: E1202 02:35:12.308053 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d85b635aeb6f72e8242f477716fbb2c27140318b1a857e55ab194a1a21146c2e\": container with ID starting with d85b635aeb6f72e8242f477716fbb2c27140318b1a857e55ab194a1a21146c2e not found: ID does not exist" containerID="d85b635aeb6f72e8242f477716fbb2c27140318b1a857e55ab194a1a21146c2e" Dec 02 02:35:12 crc kubenswrapper[4884]: I1202 02:35:12.308091 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d85b635aeb6f72e8242f477716fbb2c27140318b1a857e55ab194a1a21146c2e"} err="failed to get container status \"d85b635aeb6f72e8242f477716fbb2c27140318b1a857e55ab194a1a21146c2e\": rpc error: code = NotFound desc = could not find container \"d85b635aeb6f72e8242f477716fbb2c27140318b1a857e55ab194a1a21146c2e\": container with ID starting with d85b635aeb6f72e8242f477716fbb2c27140318b1a857e55ab194a1a21146c2e not found: ID does not exist" Dec 02 02:35:13 crc kubenswrapper[4884]: I1202 02:35:13.633091 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb115e59-faa3-454c-958c-cbeac3d80750" path="/var/lib/kubelet/pods/bb115e59-faa3-454c-958c-cbeac3d80750/volumes" Dec 02 02:35:44 crc kubenswrapper[4884]: E1202 02:35:44.079523 4884 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.32:53206->38.102.83.32:38815: write tcp 38.102.83.32:53206->38.102.83.32:38815: write: broken pipe Dec 02 02:35:46 crc kubenswrapper[4884]: I1202 02:35:46.971834 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:35:46 crc kubenswrapper[4884]: I1202 02:35:46.972687 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:36:16 crc kubenswrapper[4884]: I1202 02:36:16.971713 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:36:16 crc kubenswrapper[4884]: I1202 02:36:16.972169 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:36:46 crc kubenswrapper[4884]: I1202 02:36:46.972112 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:36:46 crc kubenswrapper[4884]: I1202 02:36:46.972841 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:36:46 crc kubenswrapper[4884]: I1202 02:36:46.972896 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 02:36:46 crc kubenswrapper[4884]: I1202 02:36:46.973673 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1171eeb0bae3a98055f356e1079ce475be63ba1276ce35ce9e5b5ec76d526801"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 02:36:46 crc kubenswrapper[4884]: I1202 02:36:46.973732 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://1171eeb0bae3a98055f356e1079ce475be63ba1276ce35ce9e5b5ec76d526801" gracePeriod=600 Dec 02 02:36:47 crc kubenswrapper[4884]: I1202 02:36:47.256648 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="1171eeb0bae3a98055f356e1079ce475be63ba1276ce35ce9e5b5ec76d526801" exitCode=0 Dec 02 02:36:47 crc kubenswrapper[4884]: I1202 02:36:47.256766 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"1171eeb0bae3a98055f356e1079ce475be63ba1276ce35ce9e5b5ec76d526801"} Dec 02 02:36:47 crc kubenswrapper[4884]: I1202 02:36:47.257142 4884 scope.go:117] "RemoveContainer" containerID="bbac8ce91b0c291a9f9471071a4f5df45f0fc7cefcbe23d4b67f1464e05c03a9" Dec 02 02:36:48 crc kubenswrapper[4884]: I1202 02:36:48.277192 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa"} Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.643211 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7t5n7"] Dec 02 02:38:15 crc kubenswrapper[4884]: E1202 02:38:15.644393 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb115e59-faa3-454c-958c-cbeac3d80750" containerName="registry-server" Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.644418 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb115e59-faa3-454c-958c-cbeac3d80750" containerName="registry-server" Dec 02 02:38:15 crc kubenswrapper[4884]: E1202 02:38:15.644458 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb115e59-faa3-454c-958c-cbeac3d80750" containerName="extract-content" Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.644470 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb115e59-faa3-454c-958c-cbeac3d80750" containerName="extract-content" Dec 02 02:38:15 crc kubenswrapper[4884]: E1202 02:38:15.644520 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb115e59-faa3-454c-958c-cbeac3d80750" containerName="extract-utilities" Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.644534 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb115e59-faa3-454c-958c-cbeac3d80750" containerName="extract-utilities" Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.644886 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb115e59-faa3-454c-958c-cbeac3d80750" containerName="registry-server" Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.647171 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.675518 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7t5n7"] Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.709084 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a2f28de-d9b9-45b8-abfe-f8e721224075-catalog-content\") pod \"redhat-operators-7t5n7\" (UID: \"5a2f28de-d9b9-45b8-abfe-f8e721224075\") " pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.709216 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a2f28de-d9b9-45b8-abfe-f8e721224075-utilities\") pod \"redhat-operators-7t5n7\" (UID: \"5a2f28de-d9b9-45b8-abfe-f8e721224075\") " pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.709278 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl99k\" (UniqueName: \"kubernetes.io/projected/5a2f28de-d9b9-45b8-abfe-f8e721224075-kube-api-access-fl99k\") pod \"redhat-operators-7t5n7\" (UID: \"5a2f28de-d9b9-45b8-abfe-f8e721224075\") " pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.810924 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a2f28de-d9b9-45b8-abfe-f8e721224075-utilities\") pod \"redhat-operators-7t5n7\" (UID: \"5a2f28de-d9b9-45b8-abfe-f8e721224075\") " pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.811002 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl99k\" (UniqueName: \"kubernetes.io/projected/5a2f28de-d9b9-45b8-abfe-f8e721224075-kube-api-access-fl99k\") pod \"redhat-operators-7t5n7\" (UID: \"5a2f28de-d9b9-45b8-abfe-f8e721224075\") " pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.811052 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a2f28de-d9b9-45b8-abfe-f8e721224075-catalog-content\") pod \"redhat-operators-7t5n7\" (UID: \"5a2f28de-d9b9-45b8-abfe-f8e721224075\") " pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.811493 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a2f28de-d9b9-45b8-abfe-f8e721224075-catalog-content\") pod \"redhat-operators-7t5n7\" (UID: \"5a2f28de-d9b9-45b8-abfe-f8e721224075\") " pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.811960 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a2f28de-d9b9-45b8-abfe-f8e721224075-utilities\") pod \"redhat-operators-7t5n7\" (UID: \"5a2f28de-d9b9-45b8-abfe-f8e721224075\") " pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.850052 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl99k\" (UniqueName: \"kubernetes.io/projected/5a2f28de-d9b9-45b8-abfe-f8e721224075-kube-api-access-fl99k\") pod \"redhat-operators-7t5n7\" (UID: \"5a2f28de-d9b9-45b8-abfe-f8e721224075\") " pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:15 crc kubenswrapper[4884]: I1202 02:38:15.980312 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:16 crc kubenswrapper[4884]: I1202 02:38:16.490204 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7t5n7"] Dec 02 02:38:16 crc kubenswrapper[4884]: W1202 02:38:16.492618 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a2f28de_d9b9_45b8_abfe_f8e721224075.slice/crio-875ec4ba5b25ac0cfcd56ae318df179b30eb0b7a61cc040dbeab89edd33874c4 WatchSource:0}: Error finding container 875ec4ba5b25ac0cfcd56ae318df179b30eb0b7a61cc040dbeab89edd33874c4: Status 404 returned error can't find the container with id 875ec4ba5b25ac0cfcd56ae318df179b30eb0b7a61cc040dbeab89edd33874c4 Dec 02 02:38:17 crc kubenswrapper[4884]: I1202 02:38:17.407522 4884 generic.go:334] "Generic (PLEG): container finished" podID="5a2f28de-d9b9-45b8-abfe-f8e721224075" containerID="13cf6f663eb46a7b95c63ec8a7a7535b09db7326877fba3e9944ad56e13946d1" exitCode=0 Dec 02 02:38:17 crc kubenswrapper[4884]: I1202 02:38:17.407606 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t5n7" event={"ID":"5a2f28de-d9b9-45b8-abfe-f8e721224075","Type":"ContainerDied","Data":"13cf6f663eb46a7b95c63ec8a7a7535b09db7326877fba3e9944ad56e13946d1"} Dec 02 02:38:17 crc kubenswrapper[4884]: I1202 02:38:17.407937 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t5n7" event={"ID":"5a2f28de-d9b9-45b8-abfe-f8e721224075","Type":"ContainerStarted","Data":"875ec4ba5b25ac0cfcd56ae318df179b30eb0b7a61cc040dbeab89edd33874c4"} Dec 02 02:38:18 crc kubenswrapper[4884]: I1202 02:38:18.421426 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t5n7" event={"ID":"5a2f28de-d9b9-45b8-abfe-f8e721224075","Type":"ContainerStarted","Data":"fa5d7e54ace0e02cf4663dc6d0b8bcd171667cd8879b71d95ca0a15fe0ed0d83"} Dec 02 02:38:20 crc kubenswrapper[4884]: I1202 02:38:20.461102 4884 generic.go:334] "Generic (PLEG): container finished" podID="5a2f28de-d9b9-45b8-abfe-f8e721224075" containerID="fa5d7e54ace0e02cf4663dc6d0b8bcd171667cd8879b71d95ca0a15fe0ed0d83" exitCode=0 Dec 02 02:38:20 crc kubenswrapper[4884]: I1202 02:38:20.461151 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t5n7" event={"ID":"5a2f28de-d9b9-45b8-abfe-f8e721224075","Type":"ContainerDied","Data":"fa5d7e54ace0e02cf4663dc6d0b8bcd171667cd8879b71d95ca0a15fe0ed0d83"} Dec 02 02:38:21 crc kubenswrapper[4884]: I1202 02:38:21.484843 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t5n7" event={"ID":"5a2f28de-d9b9-45b8-abfe-f8e721224075","Type":"ContainerStarted","Data":"dd92bcb5a230f907aba0a476e8ec968b0ee7ae300840c6bc6fc985659bf88968"} Dec 02 02:38:25 crc kubenswrapper[4884]: I1202 02:38:25.981031 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:25 crc kubenswrapper[4884]: I1202 02:38:25.981492 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:27 crc kubenswrapper[4884]: I1202 02:38:27.075799 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7t5n7" podUID="5a2f28de-d9b9-45b8-abfe-f8e721224075" containerName="registry-server" probeResult="failure" output=< Dec 02 02:38:27 crc kubenswrapper[4884]: timeout: failed to connect service ":50051" within 1s Dec 02 02:38:27 crc kubenswrapper[4884]: > Dec 02 02:38:36 crc kubenswrapper[4884]: I1202 02:38:36.070897 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:36 crc kubenswrapper[4884]: I1202 02:38:36.120220 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7t5n7" podStartSLOduration=17.466062724 podStartE2EDuration="21.120186431s" podCreationTimestamp="2025-12-02 02:38:15 +0000 UTC" firstStartedPulling="2025-12-02 02:38:17.411095129 +0000 UTC m=+3714.086932013" lastFinishedPulling="2025-12-02 02:38:21.065218826 +0000 UTC m=+3717.741055720" observedRunningTime="2025-12-02 02:38:21.505604301 +0000 UTC m=+3718.181441185" watchObservedRunningTime="2025-12-02 02:38:36.120186431 +0000 UTC m=+3732.796023345" Dec 02 02:38:36 crc kubenswrapper[4884]: I1202 02:38:36.157833 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:36 crc kubenswrapper[4884]: I1202 02:38:36.318809 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7t5n7"] Dec 02 02:38:37 crc kubenswrapper[4884]: I1202 02:38:37.666672 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7t5n7" podUID="5a2f28de-d9b9-45b8-abfe-f8e721224075" containerName="registry-server" containerID="cri-o://dd92bcb5a230f907aba0a476e8ec968b0ee7ae300840c6bc6fc985659bf88968" gracePeriod=2 Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.287126 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.313642 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a2f28de-d9b9-45b8-abfe-f8e721224075-utilities\") pod \"5a2f28de-d9b9-45b8-abfe-f8e721224075\" (UID: \"5a2f28de-d9b9-45b8-abfe-f8e721224075\") " Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.313761 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a2f28de-d9b9-45b8-abfe-f8e721224075-catalog-content\") pod \"5a2f28de-d9b9-45b8-abfe-f8e721224075\" (UID: \"5a2f28de-d9b9-45b8-abfe-f8e721224075\") " Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.313861 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fl99k\" (UniqueName: \"kubernetes.io/projected/5a2f28de-d9b9-45b8-abfe-f8e721224075-kube-api-access-fl99k\") pod \"5a2f28de-d9b9-45b8-abfe-f8e721224075\" (UID: \"5a2f28de-d9b9-45b8-abfe-f8e721224075\") " Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.314393 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a2f28de-d9b9-45b8-abfe-f8e721224075-utilities" (OuterVolumeSpecName: "utilities") pod "5a2f28de-d9b9-45b8-abfe-f8e721224075" (UID: "5a2f28de-d9b9-45b8-abfe-f8e721224075"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.322130 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a2f28de-d9b9-45b8-abfe-f8e721224075-kube-api-access-fl99k" (OuterVolumeSpecName: "kube-api-access-fl99k") pod "5a2f28de-d9b9-45b8-abfe-f8e721224075" (UID: "5a2f28de-d9b9-45b8-abfe-f8e721224075"). InnerVolumeSpecName "kube-api-access-fl99k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.416260 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fl99k\" (UniqueName: \"kubernetes.io/projected/5a2f28de-d9b9-45b8-abfe-f8e721224075-kube-api-access-fl99k\") on node \"crc\" DevicePath \"\"" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.416296 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5a2f28de-d9b9-45b8-abfe-f8e721224075-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.453821 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a2f28de-d9b9-45b8-abfe-f8e721224075-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5a2f28de-d9b9-45b8-abfe-f8e721224075" (UID: "5a2f28de-d9b9-45b8-abfe-f8e721224075"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.517850 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5a2f28de-d9b9-45b8-abfe-f8e721224075-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.679915 4884 generic.go:334] "Generic (PLEG): container finished" podID="5a2f28de-d9b9-45b8-abfe-f8e721224075" containerID="dd92bcb5a230f907aba0a476e8ec968b0ee7ae300840c6bc6fc985659bf88968" exitCode=0 Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.679973 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t5n7" event={"ID":"5a2f28de-d9b9-45b8-abfe-f8e721224075","Type":"ContainerDied","Data":"dd92bcb5a230f907aba0a476e8ec968b0ee7ae300840c6bc6fc985659bf88968"} Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.680021 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7t5n7" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.680046 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7t5n7" event={"ID":"5a2f28de-d9b9-45b8-abfe-f8e721224075","Type":"ContainerDied","Data":"875ec4ba5b25ac0cfcd56ae318df179b30eb0b7a61cc040dbeab89edd33874c4"} Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.680075 4884 scope.go:117] "RemoveContainer" containerID="dd92bcb5a230f907aba0a476e8ec968b0ee7ae300840c6bc6fc985659bf88968" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.712574 4884 scope.go:117] "RemoveContainer" containerID="fa5d7e54ace0e02cf4663dc6d0b8bcd171667cd8879b71d95ca0a15fe0ed0d83" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.743863 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7t5n7"] Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.753425 4884 scope.go:117] "RemoveContainer" containerID="13cf6f663eb46a7b95c63ec8a7a7535b09db7326877fba3e9944ad56e13946d1" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.763504 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7t5n7"] Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.799366 4884 scope.go:117] "RemoveContainer" containerID="dd92bcb5a230f907aba0a476e8ec968b0ee7ae300840c6bc6fc985659bf88968" Dec 02 02:38:38 crc kubenswrapper[4884]: E1202 02:38:38.800055 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd92bcb5a230f907aba0a476e8ec968b0ee7ae300840c6bc6fc985659bf88968\": container with ID starting with dd92bcb5a230f907aba0a476e8ec968b0ee7ae300840c6bc6fc985659bf88968 not found: ID does not exist" containerID="dd92bcb5a230f907aba0a476e8ec968b0ee7ae300840c6bc6fc985659bf88968" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.800108 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd92bcb5a230f907aba0a476e8ec968b0ee7ae300840c6bc6fc985659bf88968"} err="failed to get container status \"dd92bcb5a230f907aba0a476e8ec968b0ee7ae300840c6bc6fc985659bf88968\": rpc error: code = NotFound desc = could not find container \"dd92bcb5a230f907aba0a476e8ec968b0ee7ae300840c6bc6fc985659bf88968\": container with ID starting with dd92bcb5a230f907aba0a476e8ec968b0ee7ae300840c6bc6fc985659bf88968 not found: ID does not exist" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.800142 4884 scope.go:117] "RemoveContainer" containerID="fa5d7e54ace0e02cf4663dc6d0b8bcd171667cd8879b71d95ca0a15fe0ed0d83" Dec 02 02:38:38 crc kubenswrapper[4884]: E1202 02:38:38.800727 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa5d7e54ace0e02cf4663dc6d0b8bcd171667cd8879b71d95ca0a15fe0ed0d83\": container with ID starting with fa5d7e54ace0e02cf4663dc6d0b8bcd171667cd8879b71d95ca0a15fe0ed0d83 not found: ID does not exist" containerID="fa5d7e54ace0e02cf4663dc6d0b8bcd171667cd8879b71d95ca0a15fe0ed0d83" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.800853 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa5d7e54ace0e02cf4663dc6d0b8bcd171667cd8879b71d95ca0a15fe0ed0d83"} err="failed to get container status \"fa5d7e54ace0e02cf4663dc6d0b8bcd171667cd8879b71d95ca0a15fe0ed0d83\": rpc error: code = NotFound desc = could not find container \"fa5d7e54ace0e02cf4663dc6d0b8bcd171667cd8879b71d95ca0a15fe0ed0d83\": container with ID starting with fa5d7e54ace0e02cf4663dc6d0b8bcd171667cd8879b71d95ca0a15fe0ed0d83 not found: ID does not exist" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.800909 4884 scope.go:117] "RemoveContainer" containerID="13cf6f663eb46a7b95c63ec8a7a7535b09db7326877fba3e9944ad56e13946d1" Dec 02 02:38:38 crc kubenswrapper[4884]: E1202 02:38:38.801526 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13cf6f663eb46a7b95c63ec8a7a7535b09db7326877fba3e9944ad56e13946d1\": container with ID starting with 13cf6f663eb46a7b95c63ec8a7a7535b09db7326877fba3e9944ad56e13946d1 not found: ID does not exist" containerID="13cf6f663eb46a7b95c63ec8a7a7535b09db7326877fba3e9944ad56e13946d1" Dec 02 02:38:38 crc kubenswrapper[4884]: I1202 02:38:38.801585 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13cf6f663eb46a7b95c63ec8a7a7535b09db7326877fba3e9944ad56e13946d1"} err="failed to get container status \"13cf6f663eb46a7b95c63ec8a7a7535b09db7326877fba3e9944ad56e13946d1\": rpc error: code = NotFound desc = could not find container \"13cf6f663eb46a7b95c63ec8a7a7535b09db7326877fba3e9944ad56e13946d1\": container with ID starting with 13cf6f663eb46a7b95c63ec8a7a7535b09db7326877fba3e9944ad56e13946d1 not found: ID does not exist" Dec 02 02:38:39 crc kubenswrapper[4884]: I1202 02:38:39.627712 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a2f28de-d9b9-45b8-abfe-f8e721224075" path="/var/lib/kubelet/pods/5a2f28de-d9b9-45b8-abfe-f8e721224075/volumes" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.429516 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mdt2b"] Dec 02 02:39:04 crc kubenswrapper[4884]: E1202 02:39:04.430527 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a2f28de-d9b9-45b8-abfe-f8e721224075" containerName="extract-content" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.430542 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a2f28de-d9b9-45b8-abfe-f8e721224075" containerName="extract-content" Dec 02 02:39:04 crc kubenswrapper[4884]: E1202 02:39:04.430560 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a2f28de-d9b9-45b8-abfe-f8e721224075" containerName="extract-utilities" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.430566 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a2f28de-d9b9-45b8-abfe-f8e721224075" containerName="extract-utilities" Dec 02 02:39:04 crc kubenswrapper[4884]: E1202 02:39:04.430584 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a2f28de-d9b9-45b8-abfe-f8e721224075" containerName="registry-server" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.430589 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a2f28de-d9b9-45b8-abfe-f8e721224075" containerName="registry-server" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.430779 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a2f28de-d9b9-45b8-abfe-f8e721224075" containerName="registry-server" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.432400 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.452452 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mdt2b"] Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.589827 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld4zl\" (UniqueName: \"kubernetes.io/projected/d77c7d42-c974-4387-b041-cc5ccbe934c4-kube-api-access-ld4zl\") pod \"certified-operators-mdt2b\" (UID: \"d77c7d42-c974-4387-b041-cc5ccbe934c4\") " pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.589874 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d77c7d42-c974-4387-b041-cc5ccbe934c4-catalog-content\") pod \"certified-operators-mdt2b\" (UID: \"d77c7d42-c974-4387-b041-cc5ccbe934c4\") " pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.590133 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d77c7d42-c974-4387-b041-cc5ccbe934c4-utilities\") pod \"certified-operators-mdt2b\" (UID: \"d77c7d42-c974-4387-b041-cc5ccbe934c4\") " pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.692043 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld4zl\" (UniqueName: \"kubernetes.io/projected/d77c7d42-c974-4387-b041-cc5ccbe934c4-kube-api-access-ld4zl\") pod \"certified-operators-mdt2b\" (UID: \"d77c7d42-c974-4387-b041-cc5ccbe934c4\") " pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.692114 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d77c7d42-c974-4387-b041-cc5ccbe934c4-catalog-content\") pod \"certified-operators-mdt2b\" (UID: \"d77c7d42-c974-4387-b041-cc5ccbe934c4\") " pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.692228 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d77c7d42-c974-4387-b041-cc5ccbe934c4-utilities\") pod \"certified-operators-mdt2b\" (UID: \"d77c7d42-c974-4387-b041-cc5ccbe934c4\") " pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.692582 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d77c7d42-c974-4387-b041-cc5ccbe934c4-catalog-content\") pod \"certified-operators-mdt2b\" (UID: \"d77c7d42-c974-4387-b041-cc5ccbe934c4\") " pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.692647 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d77c7d42-c974-4387-b041-cc5ccbe934c4-utilities\") pod \"certified-operators-mdt2b\" (UID: \"d77c7d42-c974-4387-b041-cc5ccbe934c4\") " pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.713796 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld4zl\" (UniqueName: \"kubernetes.io/projected/d77c7d42-c974-4387-b041-cc5ccbe934c4-kube-api-access-ld4zl\") pod \"certified-operators-mdt2b\" (UID: \"d77c7d42-c974-4387-b041-cc5ccbe934c4\") " pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:04 crc kubenswrapper[4884]: I1202 02:39:04.759642 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:05 crc kubenswrapper[4884]: I1202 02:39:05.304091 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mdt2b"] Dec 02 02:39:05 crc kubenswrapper[4884]: I1202 02:39:05.996249 4884 generic.go:334] "Generic (PLEG): container finished" podID="d77c7d42-c974-4387-b041-cc5ccbe934c4" containerID="19d3d2d7d6c7f995a57af2152ad2f059f8a3691d56410d0b5d9f3e05d5d801e7" exitCode=0 Dec 02 02:39:05 crc kubenswrapper[4884]: I1202 02:39:05.996306 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mdt2b" event={"ID":"d77c7d42-c974-4387-b041-cc5ccbe934c4","Type":"ContainerDied","Data":"19d3d2d7d6c7f995a57af2152ad2f059f8a3691d56410d0b5d9f3e05d5d801e7"} Dec 02 02:39:05 crc kubenswrapper[4884]: I1202 02:39:05.996642 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mdt2b" event={"ID":"d77c7d42-c974-4387-b041-cc5ccbe934c4","Type":"ContainerStarted","Data":"783b9a5e3731f0296da9b7bf362e39ab85f1fc346d3b3e921aa66a7de0377260"} Dec 02 02:39:08 crc kubenswrapper[4884]: I1202 02:39:08.019998 4884 generic.go:334] "Generic (PLEG): container finished" podID="d77c7d42-c974-4387-b041-cc5ccbe934c4" containerID="21564d8fc3a17907f72642425faf2160fb2debb81433935a4d7851ed4e519be8" exitCode=0 Dec 02 02:39:08 crc kubenswrapper[4884]: I1202 02:39:08.020128 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mdt2b" event={"ID":"d77c7d42-c974-4387-b041-cc5ccbe934c4","Type":"ContainerDied","Data":"21564d8fc3a17907f72642425faf2160fb2debb81433935a4d7851ed4e519be8"} Dec 02 02:39:09 crc kubenswrapper[4884]: I1202 02:39:09.036384 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mdt2b" event={"ID":"d77c7d42-c974-4387-b041-cc5ccbe934c4","Type":"ContainerStarted","Data":"5af7c39bde6698ce7084b834647b0843610e768cec26fb8e8fbc0040cabca3a4"} Dec 02 02:39:09 crc kubenswrapper[4884]: I1202 02:39:09.073179 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mdt2b" podStartSLOduration=2.54958944 podStartE2EDuration="5.07315575s" podCreationTimestamp="2025-12-02 02:39:04 +0000 UTC" firstStartedPulling="2025-12-02 02:39:05.999213156 +0000 UTC m=+3762.675050080" lastFinishedPulling="2025-12-02 02:39:08.522779466 +0000 UTC m=+3765.198616390" observedRunningTime="2025-12-02 02:39:09.059827134 +0000 UTC m=+3765.735664028" watchObservedRunningTime="2025-12-02 02:39:09.07315575 +0000 UTC m=+3765.748992644" Dec 02 02:39:14 crc kubenswrapper[4884]: I1202 02:39:14.760004 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:14 crc kubenswrapper[4884]: I1202 02:39:14.760533 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:14 crc kubenswrapper[4884]: I1202 02:39:14.838787 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:15 crc kubenswrapper[4884]: I1202 02:39:15.231376 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:15 crc kubenswrapper[4884]: I1202 02:39:15.306406 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mdt2b"] Dec 02 02:39:16 crc kubenswrapper[4884]: I1202 02:39:16.971641 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:39:16 crc kubenswrapper[4884]: I1202 02:39:16.972428 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:39:17 crc kubenswrapper[4884]: I1202 02:39:17.170981 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mdt2b" podUID="d77c7d42-c974-4387-b041-cc5ccbe934c4" containerName="registry-server" containerID="cri-o://5af7c39bde6698ce7084b834647b0843610e768cec26fb8e8fbc0040cabca3a4" gracePeriod=2 Dec 02 02:39:18 crc kubenswrapper[4884]: I1202 02:39:18.186796 4884 generic.go:334] "Generic (PLEG): container finished" podID="d77c7d42-c974-4387-b041-cc5ccbe934c4" containerID="5af7c39bde6698ce7084b834647b0843610e768cec26fb8e8fbc0040cabca3a4" exitCode=0 Dec 02 02:39:18 crc kubenswrapper[4884]: I1202 02:39:18.186886 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mdt2b" event={"ID":"d77c7d42-c974-4387-b041-cc5ccbe934c4","Type":"ContainerDied","Data":"5af7c39bde6698ce7084b834647b0843610e768cec26fb8e8fbc0040cabca3a4"} Dec 02 02:39:18 crc kubenswrapper[4884]: I1202 02:39:18.187167 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mdt2b" event={"ID":"d77c7d42-c974-4387-b041-cc5ccbe934c4","Type":"ContainerDied","Data":"783b9a5e3731f0296da9b7bf362e39ab85f1fc346d3b3e921aa66a7de0377260"} Dec 02 02:39:18 crc kubenswrapper[4884]: I1202 02:39:18.187185 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="783b9a5e3731f0296da9b7bf362e39ab85f1fc346d3b3e921aa66a7de0377260" Dec 02 02:39:18 crc kubenswrapper[4884]: I1202 02:39:18.269385 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:18 crc kubenswrapper[4884]: I1202 02:39:18.429614 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d77c7d42-c974-4387-b041-cc5ccbe934c4-catalog-content\") pod \"d77c7d42-c974-4387-b041-cc5ccbe934c4\" (UID: \"d77c7d42-c974-4387-b041-cc5ccbe934c4\") " Dec 02 02:39:18 crc kubenswrapper[4884]: I1202 02:39:18.430031 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d77c7d42-c974-4387-b041-cc5ccbe934c4-utilities\") pod \"d77c7d42-c974-4387-b041-cc5ccbe934c4\" (UID: \"d77c7d42-c974-4387-b041-cc5ccbe934c4\") " Dec 02 02:39:18 crc kubenswrapper[4884]: I1202 02:39:18.431111 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld4zl\" (UniqueName: \"kubernetes.io/projected/d77c7d42-c974-4387-b041-cc5ccbe934c4-kube-api-access-ld4zl\") pod \"d77c7d42-c974-4387-b041-cc5ccbe934c4\" (UID: \"d77c7d42-c974-4387-b041-cc5ccbe934c4\") " Dec 02 02:39:18 crc kubenswrapper[4884]: I1202 02:39:18.431452 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d77c7d42-c974-4387-b041-cc5ccbe934c4-utilities" (OuterVolumeSpecName: "utilities") pod "d77c7d42-c974-4387-b041-cc5ccbe934c4" (UID: "d77c7d42-c974-4387-b041-cc5ccbe934c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:39:18 crc kubenswrapper[4884]: I1202 02:39:18.432451 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d77c7d42-c974-4387-b041-cc5ccbe934c4-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:39:18 crc kubenswrapper[4884]: I1202 02:39:18.444485 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d77c7d42-c974-4387-b041-cc5ccbe934c4-kube-api-access-ld4zl" (OuterVolumeSpecName: "kube-api-access-ld4zl") pod "d77c7d42-c974-4387-b041-cc5ccbe934c4" (UID: "d77c7d42-c974-4387-b041-cc5ccbe934c4"). InnerVolumeSpecName "kube-api-access-ld4zl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:39:18 crc kubenswrapper[4884]: I1202 02:39:18.501415 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d77c7d42-c974-4387-b041-cc5ccbe934c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d77c7d42-c974-4387-b041-cc5ccbe934c4" (UID: "d77c7d42-c974-4387-b041-cc5ccbe934c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:39:18 crc kubenswrapper[4884]: I1202 02:39:18.535043 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld4zl\" (UniqueName: \"kubernetes.io/projected/d77c7d42-c974-4387-b041-cc5ccbe934c4-kube-api-access-ld4zl\") on node \"crc\" DevicePath \"\"" Dec 02 02:39:18 crc kubenswrapper[4884]: I1202 02:39:18.535089 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d77c7d42-c974-4387-b041-cc5ccbe934c4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:39:19 crc kubenswrapper[4884]: I1202 02:39:19.197639 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mdt2b" Dec 02 02:39:19 crc kubenswrapper[4884]: I1202 02:39:19.249584 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mdt2b"] Dec 02 02:39:19 crc kubenswrapper[4884]: I1202 02:39:19.265566 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mdt2b"] Dec 02 02:39:19 crc kubenswrapper[4884]: I1202 02:39:19.652824 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d77c7d42-c974-4387-b041-cc5ccbe934c4" path="/var/lib/kubelet/pods/d77c7d42-c974-4387-b041-cc5ccbe934c4/volumes" Dec 02 02:39:46 crc kubenswrapper[4884]: I1202 02:39:46.971264 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:39:46 crc kubenswrapper[4884]: I1202 02:39:46.971954 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:40:16 crc kubenswrapper[4884]: I1202 02:40:16.971921 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:40:16 crc kubenswrapper[4884]: I1202 02:40:16.972353 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:40:16 crc kubenswrapper[4884]: I1202 02:40:16.972399 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 02:40:16 crc kubenswrapper[4884]: I1202 02:40:16.973117 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 02:40:16 crc kubenswrapper[4884]: I1202 02:40:16.973160 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" gracePeriod=600 Dec 02 02:40:17 crc kubenswrapper[4884]: E1202 02:40:17.113701 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:40:17 crc kubenswrapper[4884]: I1202 02:40:17.941135 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" exitCode=0 Dec 02 02:40:17 crc kubenswrapper[4884]: I1202 02:40:17.941334 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa"} Dec 02 02:40:17 crc kubenswrapper[4884]: I1202 02:40:17.941542 4884 scope.go:117] "RemoveContainer" containerID="1171eeb0bae3a98055f356e1079ce475be63ba1276ce35ce9e5b5ec76d526801" Dec 02 02:40:17 crc kubenswrapper[4884]: I1202 02:40:17.942340 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:40:17 crc kubenswrapper[4884]: E1202 02:40:17.942817 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:40:29 crc kubenswrapper[4884]: I1202 02:40:29.615838 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:40:29 crc kubenswrapper[4884]: E1202 02:40:29.616454 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:40:42 crc kubenswrapper[4884]: I1202 02:40:42.614022 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:40:42 crc kubenswrapper[4884]: E1202 02:40:42.615111 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:40:55 crc kubenswrapper[4884]: I1202 02:40:55.614708 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:40:55 crc kubenswrapper[4884]: E1202 02:40:55.616000 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:41:08 crc kubenswrapper[4884]: I1202 02:41:08.614871 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:41:08 crc kubenswrapper[4884]: E1202 02:41:08.615938 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:41:23 crc kubenswrapper[4884]: I1202 02:41:23.620258 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:41:23 crc kubenswrapper[4884]: E1202 02:41:23.620903 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:41:35 crc kubenswrapper[4884]: I1202 02:41:35.615450 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:41:35 crc kubenswrapper[4884]: E1202 02:41:35.616547 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:41:48 crc kubenswrapper[4884]: I1202 02:41:48.614505 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:41:48 crc kubenswrapper[4884]: E1202 02:41:48.615855 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:42:02 crc kubenswrapper[4884]: I1202 02:42:02.614561 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:42:02 crc kubenswrapper[4884]: E1202 02:42:02.615631 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:42:15 crc kubenswrapper[4884]: I1202 02:42:15.615022 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:42:15 crc kubenswrapper[4884]: E1202 02:42:15.616196 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:42:30 crc kubenswrapper[4884]: I1202 02:42:30.615150 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:42:30 crc kubenswrapper[4884]: E1202 02:42:30.616200 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:42:42 crc kubenswrapper[4884]: I1202 02:42:42.615560 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:42:42 crc kubenswrapper[4884]: E1202 02:42:42.616523 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:42:53 crc kubenswrapper[4884]: I1202 02:42:53.623731 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:42:53 crc kubenswrapper[4884]: E1202 02:42:53.624538 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:43:04 crc kubenswrapper[4884]: I1202 02:43:04.614292 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:43:04 crc kubenswrapper[4884]: E1202 02:43:04.616172 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:43:16 crc kubenswrapper[4884]: I1202 02:43:16.614268 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:43:16 crc kubenswrapper[4884]: E1202 02:43:16.615249 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:43:27 crc kubenswrapper[4884]: I1202 02:43:27.624274 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:43:27 crc kubenswrapper[4884]: E1202 02:43:27.625445 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:43:41 crc kubenswrapper[4884]: I1202 02:43:41.615379 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:43:41 crc kubenswrapper[4884]: E1202 02:43:41.616684 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:43:54 crc kubenswrapper[4884]: I1202 02:43:54.615008 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:43:54 crc kubenswrapper[4884]: E1202 02:43:54.616337 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:44:08 crc kubenswrapper[4884]: I1202 02:44:08.614487 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:44:08 crc kubenswrapper[4884]: E1202 02:44:08.615792 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:44:23 crc kubenswrapper[4884]: I1202 02:44:23.633094 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:44:23 crc kubenswrapper[4884]: E1202 02:44:23.634628 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:44:37 crc kubenswrapper[4884]: I1202 02:44:37.620857 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:44:37 crc kubenswrapper[4884]: E1202 02:44:37.621918 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:44:49 crc kubenswrapper[4884]: I1202 02:44:49.620589 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:44:49 crc kubenswrapper[4884]: E1202 02:44:49.622141 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.191812 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f"] Dec 02 02:45:00 crc kubenswrapper[4884]: E1202 02:45:00.192701 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d77c7d42-c974-4387-b041-cc5ccbe934c4" containerName="extract-utilities" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.192714 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d77c7d42-c974-4387-b041-cc5ccbe934c4" containerName="extract-utilities" Dec 02 02:45:00 crc kubenswrapper[4884]: E1202 02:45:00.192730 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d77c7d42-c974-4387-b041-cc5ccbe934c4" containerName="extract-content" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.192736 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d77c7d42-c974-4387-b041-cc5ccbe934c4" containerName="extract-content" Dec 02 02:45:00 crc kubenswrapper[4884]: E1202 02:45:00.192786 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d77c7d42-c974-4387-b041-cc5ccbe934c4" containerName="registry-server" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.192795 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="d77c7d42-c974-4387-b041-cc5ccbe934c4" containerName="registry-server" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.192987 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="d77c7d42-c974-4387-b041-cc5ccbe934c4" containerName="registry-server" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.193602 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.195984 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.197312 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.216997 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f"] Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.237030 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-config-volume\") pod \"collect-profiles-29410725-pw29f\" (UID: \"04cffad4-0c7e-4ace-a061-8043d7c4d4e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.237087 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-secret-volume\") pod \"collect-profiles-29410725-pw29f\" (UID: \"04cffad4-0c7e-4ace-a061-8043d7c4d4e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.237181 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxgr2\" (UniqueName: \"kubernetes.io/projected/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-kube-api-access-kxgr2\") pod \"collect-profiles-29410725-pw29f\" (UID: \"04cffad4-0c7e-4ace-a061-8043d7c4d4e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.339379 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-secret-volume\") pod \"collect-profiles-29410725-pw29f\" (UID: \"04cffad4-0c7e-4ace-a061-8043d7c4d4e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.339482 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxgr2\" (UniqueName: \"kubernetes.io/projected/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-kube-api-access-kxgr2\") pod \"collect-profiles-29410725-pw29f\" (UID: \"04cffad4-0c7e-4ace-a061-8043d7c4d4e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.339592 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-config-volume\") pod \"collect-profiles-29410725-pw29f\" (UID: \"04cffad4-0c7e-4ace-a061-8043d7c4d4e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.340287 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-config-volume\") pod \"collect-profiles-29410725-pw29f\" (UID: \"04cffad4-0c7e-4ace-a061-8043d7c4d4e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.344734 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-secret-volume\") pod \"collect-profiles-29410725-pw29f\" (UID: \"04cffad4-0c7e-4ace-a061-8043d7c4d4e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.357434 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxgr2\" (UniqueName: \"kubernetes.io/projected/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-kube-api-access-kxgr2\") pod \"collect-profiles-29410725-pw29f\" (UID: \"04cffad4-0c7e-4ace-a061-8043d7c4d4e9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.523895 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" Dec 02 02:45:00 crc kubenswrapper[4884]: I1202 02:45:00.614762 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:45:00 crc kubenswrapper[4884]: E1202 02:45:00.615146 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:45:01 crc kubenswrapper[4884]: I1202 02:45:01.503866 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f"] Dec 02 02:45:02 crc kubenswrapper[4884]: I1202 02:45:02.397959 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" event={"ID":"04cffad4-0c7e-4ace-a061-8043d7c4d4e9","Type":"ContainerStarted","Data":"0ce678a0b41db2b771389315e9c303de87381cb5b016c8765c3c66acce10e8c6"} Dec 02 02:45:02 crc kubenswrapper[4884]: I1202 02:45:02.398484 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" event={"ID":"04cffad4-0c7e-4ace-a061-8043d7c4d4e9","Type":"ContainerStarted","Data":"7d5b7145c9284908831e08c3c8d0f71966f16298899b09fbed98c17e3c80e5ef"} Dec 02 02:45:02 crc kubenswrapper[4884]: I1202 02:45:02.420461 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" podStartSLOduration=2.420444788 podStartE2EDuration="2.420444788s" podCreationTimestamp="2025-12-02 02:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 02:45:02.416471311 +0000 UTC m=+4119.092308205" watchObservedRunningTime="2025-12-02 02:45:02.420444788 +0000 UTC m=+4119.096281672" Dec 02 02:45:03 crc kubenswrapper[4884]: I1202 02:45:03.411584 4884 generic.go:334] "Generic (PLEG): container finished" podID="04cffad4-0c7e-4ace-a061-8043d7c4d4e9" containerID="0ce678a0b41db2b771389315e9c303de87381cb5b016c8765c3c66acce10e8c6" exitCode=0 Dec 02 02:45:03 crc kubenswrapper[4884]: I1202 02:45:03.411695 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" event={"ID":"04cffad4-0c7e-4ace-a061-8043d7c4d4e9","Type":"ContainerDied","Data":"0ce678a0b41db2b771389315e9c303de87381cb5b016c8765c3c66acce10e8c6"} Dec 02 02:45:04 crc kubenswrapper[4884]: I1202 02:45:04.870195 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" Dec 02 02:45:05 crc kubenswrapper[4884]: I1202 02:45:05.044085 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxgr2\" (UniqueName: \"kubernetes.io/projected/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-kube-api-access-kxgr2\") pod \"04cffad4-0c7e-4ace-a061-8043d7c4d4e9\" (UID: \"04cffad4-0c7e-4ace-a061-8043d7c4d4e9\") " Dec 02 02:45:05 crc kubenswrapper[4884]: I1202 02:45:05.044243 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-config-volume\") pod \"04cffad4-0c7e-4ace-a061-8043d7c4d4e9\" (UID: \"04cffad4-0c7e-4ace-a061-8043d7c4d4e9\") " Dec 02 02:45:05 crc kubenswrapper[4884]: I1202 02:45:05.044306 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-secret-volume\") pod \"04cffad4-0c7e-4ace-a061-8043d7c4d4e9\" (UID: \"04cffad4-0c7e-4ace-a061-8043d7c4d4e9\") " Dec 02 02:45:05 crc kubenswrapper[4884]: I1202 02:45:05.044921 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-config-volume" (OuterVolumeSpecName: "config-volume") pod "04cffad4-0c7e-4ace-a061-8043d7c4d4e9" (UID: "04cffad4-0c7e-4ace-a061-8043d7c4d4e9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 02:45:05 crc kubenswrapper[4884]: I1202 02:45:05.045129 4884 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 02:45:05 crc kubenswrapper[4884]: I1202 02:45:05.051237 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-kube-api-access-kxgr2" (OuterVolumeSpecName: "kube-api-access-kxgr2") pod "04cffad4-0c7e-4ace-a061-8043d7c4d4e9" (UID: "04cffad4-0c7e-4ace-a061-8043d7c4d4e9"). InnerVolumeSpecName "kube-api-access-kxgr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:45:05 crc kubenswrapper[4884]: I1202 02:45:05.051648 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "04cffad4-0c7e-4ace-a061-8043d7c4d4e9" (UID: "04cffad4-0c7e-4ace-a061-8043d7c4d4e9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 02:45:05 crc kubenswrapper[4884]: I1202 02:45:05.147770 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxgr2\" (UniqueName: \"kubernetes.io/projected/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-kube-api-access-kxgr2\") on node \"crc\" DevicePath \"\"" Dec 02 02:45:05 crc kubenswrapper[4884]: I1202 02:45:05.147818 4884 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/04cffad4-0c7e-4ace-a061-8043d7c4d4e9-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 02:45:05 crc kubenswrapper[4884]: I1202 02:45:05.448181 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" event={"ID":"04cffad4-0c7e-4ace-a061-8043d7c4d4e9","Type":"ContainerDied","Data":"7d5b7145c9284908831e08c3c8d0f71966f16298899b09fbed98c17e3c80e5ef"} Dec 02 02:45:05 crc kubenswrapper[4884]: I1202 02:45:05.448241 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d5b7145c9284908831e08c3c8d0f71966f16298899b09fbed98c17e3c80e5ef" Dec 02 02:45:05 crc kubenswrapper[4884]: I1202 02:45:05.448274 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410725-pw29f" Dec 02 02:45:05 crc kubenswrapper[4884]: I1202 02:45:05.941953 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn"] Dec 02 02:45:05 crc kubenswrapper[4884]: I1202 02:45:05.958512 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410680-qmndn"] Dec 02 02:45:07 crc kubenswrapper[4884]: I1202 02:45:07.628151 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5869070c-4b43-4ecc-b6a0-971cd0686284" path="/var/lib/kubelet/pods/5869070c-4b43-4ecc-b6a0-971cd0686284/volumes" Dec 02 02:45:15 crc kubenswrapper[4884]: I1202 02:45:15.614835 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:45:15 crc kubenswrapper[4884]: E1202 02:45:15.616066 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:45:28 crc kubenswrapper[4884]: I1202 02:45:28.613769 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:45:29 crc kubenswrapper[4884]: I1202 02:45:29.761696 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"579c75b60c975cce5b8ce8cb9ee141c65e4945b53bab8840e41aec5e8d052b1c"} Dec 02 02:45:46 crc kubenswrapper[4884]: I1202 02:45:46.754333 4884 scope.go:117] "RemoveContainer" containerID="5af7c39bde6698ce7084b834647b0843610e768cec26fb8e8fbc0040cabca3a4" Dec 02 02:45:46 crc kubenswrapper[4884]: I1202 02:45:46.783016 4884 scope.go:117] "RemoveContainer" containerID="19d3d2d7d6c7f995a57af2152ad2f059f8a3691d56410d0b5d9f3e05d5d801e7" Dec 02 02:45:46 crc kubenswrapper[4884]: I1202 02:45:46.822488 4884 scope.go:117] "RemoveContainer" containerID="ce10549b269f041591f6b05a9a730ea509a7dbdeeb9780199b39e6516401a936" Dec 02 02:45:46 crc kubenswrapper[4884]: I1202 02:45:46.887181 4884 scope.go:117] "RemoveContainer" containerID="21564d8fc3a17907f72642425faf2160fb2debb81433935a4d7851ed4e519be8" Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.006673 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f6ph5"] Dec 02 02:45:53 crc kubenswrapper[4884]: E1202 02:45:53.008111 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04cffad4-0c7e-4ace-a061-8043d7c4d4e9" containerName="collect-profiles" Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.008138 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="04cffad4-0c7e-4ace-a061-8043d7c4d4e9" containerName="collect-profiles" Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.008583 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="04cffad4-0c7e-4ace-a061-8043d7c4d4e9" containerName="collect-profiles" Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.011676 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.024839 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f6ph5"] Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.101739 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6psxj\" (UniqueName: \"kubernetes.io/projected/791205ea-a0c8-476a-827d-5cbc1a24e774-kube-api-access-6psxj\") pod \"community-operators-f6ph5\" (UID: \"791205ea-a0c8-476a-827d-5cbc1a24e774\") " pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.102098 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/791205ea-a0c8-476a-827d-5cbc1a24e774-catalog-content\") pod \"community-operators-f6ph5\" (UID: \"791205ea-a0c8-476a-827d-5cbc1a24e774\") " pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.102212 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/791205ea-a0c8-476a-827d-5cbc1a24e774-utilities\") pod \"community-operators-f6ph5\" (UID: \"791205ea-a0c8-476a-827d-5cbc1a24e774\") " pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.204783 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6psxj\" (UniqueName: \"kubernetes.io/projected/791205ea-a0c8-476a-827d-5cbc1a24e774-kube-api-access-6psxj\") pod \"community-operators-f6ph5\" (UID: \"791205ea-a0c8-476a-827d-5cbc1a24e774\") " pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.204939 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/791205ea-a0c8-476a-827d-5cbc1a24e774-catalog-content\") pod \"community-operators-f6ph5\" (UID: \"791205ea-a0c8-476a-827d-5cbc1a24e774\") " pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.205119 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/791205ea-a0c8-476a-827d-5cbc1a24e774-utilities\") pod \"community-operators-f6ph5\" (UID: \"791205ea-a0c8-476a-827d-5cbc1a24e774\") " pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.205472 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/791205ea-a0c8-476a-827d-5cbc1a24e774-catalog-content\") pod \"community-operators-f6ph5\" (UID: \"791205ea-a0c8-476a-827d-5cbc1a24e774\") " pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.205552 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/791205ea-a0c8-476a-827d-5cbc1a24e774-utilities\") pod \"community-operators-f6ph5\" (UID: \"791205ea-a0c8-476a-827d-5cbc1a24e774\") " pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.233885 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6psxj\" (UniqueName: \"kubernetes.io/projected/791205ea-a0c8-476a-827d-5cbc1a24e774-kube-api-access-6psxj\") pod \"community-operators-f6ph5\" (UID: \"791205ea-a0c8-476a-827d-5cbc1a24e774\") " pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.337347 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:45:53 crc kubenswrapper[4884]: I1202 02:45:53.935595 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f6ph5"] Dec 02 02:45:53 crc kubenswrapper[4884]: W1202 02:45:53.940924 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod791205ea_a0c8_476a_827d_5cbc1a24e774.slice/crio-baa660c49fc663da5411fdaab902ddd94ba3e42a11b3b0e999807feeb702684b WatchSource:0}: Error finding container baa660c49fc663da5411fdaab902ddd94ba3e42a11b3b0e999807feeb702684b: Status 404 returned error can't find the container with id baa660c49fc663da5411fdaab902ddd94ba3e42a11b3b0e999807feeb702684b Dec 02 02:45:54 crc kubenswrapper[4884]: I1202 02:45:54.049368 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f6ph5" event={"ID":"791205ea-a0c8-476a-827d-5cbc1a24e774","Type":"ContainerStarted","Data":"baa660c49fc663da5411fdaab902ddd94ba3e42a11b3b0e999807feeb702684b"} Dec 02 02:45:55 crc kubenswrapper[4884]: I1202 02:45:55.064189 4884 generic.go:334] "Generic (PLEG): container finished" podID="791205ea-a0c8-476a-827d-5cbc1a24e774" containerID="df3fbc0e0ad84b1d7f090bec27bd4b18339c28e99e21d11041a55dd565cb918d" exitCode=0 Dec 02 02:45:55 crc kubenswrapper[4884]: I1202 02:45:55.064333 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f6ph5" event={"ID":"791205ea-a0c8-476a-827d-5cbc1a24e774","Type":"ContainerDied","Data":"df3fbc0e0ad84b1d7f090bec27bd4b18339c28e99e21d11041a55dd565cb918d"} Dec 02 02:45:55 crc kubenswrapper[4884]: I1202 02:45:55.067550 4884 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 02:45:57 crc kubenswrapper[4884]: I1202 02:45:57.087660 4884 generic.go:334] "Generic (PLEG): container finished" podID="791205ea-a0c8-476a-827d-5cbc1a24e774" containerID="e47bc953828954ab2f4f5e5b7dbb8deb44cfc3cf75740d0c9df409067e1c80a7" exitCode=0 Dec 02 02:45:57 crc kubenswrapper[4884]: I1202 02:45:57.087912 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f6ph5" event={"ID":"791205ea-a0c8-476a-827d-5cbc1a24e774","Type":"ContainerDied","Data":"e47bc953828954ab2f4f5e5b7dbb8deb44cfc3cf75740d0c9df409067e1c80a7"} Dec 02 02:45:58 crc kubenswrapper[4884]: I1202 02:45:58.099611 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f6ph5" event={"ID":"791205ea-a0c8-476a-827d-5cbc1a24e774","Type":"ContainerStarted","Data":"1efcf5ffe8b307358ba20f5a5c0dd1d3beaac822091bb6ec447e6bc32003f950"} Dec 02 02:45:58 crc kubenswrapper[4884]: I1202 02:45:58.128949 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f6ph5" podStartSLOduration=3.5712110089999998 podStartE2EDuration="6.128928649s" podCreationTimestamp="2025-12-02 02:45:52 +0000 UTC" firstStartedPulling="2025-12-02 02:45:55.067145937 +0000 UTC m=+4171.742982851" lastFinishedPulling="2025-12-02 02:45:57.624863597 +0000 UTC m=+4174.300700491" observedRunningTime="2025-12-02 02:45:58.113243535 +0000 UTC m=+4174.789080419" watchObservedRunningTime="2025-12-02 02:45:58.128928649 +0000 UTC m=+4174.804765543" Dec 02 02:46:03 crc kubenswrapper[4884]: I1202 02:46:03.338344 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:46:03 crc kubenswrapper[4884]: I1202 02:46:03.338967 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:46:03 crc kubenswrapper[4884]: I1202 02:46:03.423918 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:46:04 crc kubenswrapper[4884]: I1202 02:46:04.230453 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:46:04 crc kubenswrapper[4884]: I1202 02:46:04.291479 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f6ph5"] Dec 02 02:46:06 crc kubenswrapper[4884]: I1202 02:46:06.182049 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f6ph5" podUID="791205ea-a0c8-476a-827d-5cbc1a24e774" containerName="registry-server" containerID="cri-o://1efcf5ffe8b307358ba20f5a5c0dd1d3beaac822091bb6ec447e6bc32003f950" gracePeriod=2 Dec 02 02:46:06 crc kubenswrapper[4884]: I1202 02:46:06.695063 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:46:06 crc kubenswrapper[4884]: I1202 02:46:06.741577 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6psxj\" (UniqueName: \"kubernetes.io/projected/791205ea-a0c8-476a-827d-5cbc1a24e774-kube-api-access-6psxj\") pod \"791205ea-a0c8-476a-827d-5cbc1a24e774\" (UID: \"791205ea-a0c8-476a-827d-5cbc1a24e774\") " Dec 02 02:46:06 crc kubenswrapper[4884]: I1202 02:46:06.741664 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/791205ea-a0c8-476a-827d-5cbc1a24e774-catalog-content\") pod \"791205ea-a0c8-476a-827d-5cbc1a24e774\" (UID: \"791205ea-a0c8-476a-827d-5cbc1a24e774\") " Dec 02 02:46:06 crc kubenswrapper[4884]: I1202 02:46:06.741899 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/791205ea-a0c8-476a-827d-5cbc1a24e774-utilities\") pod \"791205ea-a0c8-476a-827d-5cbc1a24e774\" (UID: \"791205ea-a0c8-476a-827d-5cbc1a24e774\") " Dec 02 02:46:06 crc kubenswrapper[4884]: I1202 02:46:06.743285 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/791205ea-a0c8-476a-827d-5cbc1a24e774-utilities" (OuterVolumeSpecName: "utilities") pod "791205ea-a0c8-476a-827d-5cbc1a24e774" (UID: "791205ea-a0c8-476a-827d-5cbc1a24e774"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:46:06 crc kubenswrapper[4884]: I1202 02:46:06.751402 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/791205ea-a0c8-476a-827d-5cbc1a24e774-kube-api-access-6psxj" (OuterVolumeSpecName: "kube-api-access-6psxj") pod "791205ea-a0c8-476a-827d-5cbc1a24e774" (UID: "791205ea-a0c8-476a-827d-5cbc1a24e774"). InnerVolumeSpecName "kube-api-access-6psxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:46:06 crc kubenswrapper[4884]: I1202 02:46:06.819065 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/791205ea-a0c8-476a-827d-5cbc1a24e774-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "791205ea-a0c8-476a-827d-5cbc1a24e774" (UID: "791205ea-a0c8-476a-827d-5cbc1a24e774"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:46:06 crc kubenswrapper[4884]: I1202 02:46:06.843407 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6psxj\" (UniqueName: \"kubernetes.io/projected/791205ea-a0c8-476a-827d-5cbc1a24e774-kube-api-access-6psxj\") on node \"crc\" DevicePath \"\"" Dec 02 02:46:06 crc kubenswrapper[4884]: I1202 02:46:06.843443 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/791205ea-a0c8-476a-827d-5cbc1a24e774-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:46:06 crc kubenswrapper[4884]: I1202 02:46:06.843457 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/791205ea-a0c8-476a-827d-5cbc1a24e774-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.198503 4884 generic.go:334] "Generic (PLEG): container finished" podID="791205ea-a0c8-476a-827d-5cbc1a24e774" containerID="1efcf5ffe8b307358ba20f5a5c0dd1d3beaac822091bb6ec447e6bc32003f950" exitCode=0 Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.198581 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f6ph5" Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.198648 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f6ph5" event={"ID":"791205ea-a0c8-476a-827d-5cbc1a24e774","Type":"ContainerDied","Data":"1efcf5ffe8b307358ba20f5a5c0dd1d3beaac822091bb6ec447e6bc32003f950"} Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.199377 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f6ph5" event={"ID":"791205ea-a0c8-476a-827d-5cbc1a24e774","Type":"ContainerDied","Data":"baa660c49fc663da5411fdaab902ddd94ba3e42a11b3b0e999807feeb702684b"} Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.199418 4884 scope.go:117] "RemoveContainer" containerID="1efcf5ffe8b307358ba20f5a5c0dd1d3beaac822091bb6ec447e6bc32003f950" Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.247320 4884 scope.go:117] "RemoveContainer" containerID="e47bc953828954ab2f4f5e5b7dbb8deb44cfc3cf75740d0c9df409067e1c80a7" Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.276969 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f6ph5"] Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.289860 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f6ph5"] Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.290210 4884 scope.go:117] "RemoveContainer" containerID="df3fbc0e0ad84b1d7f090bec27bd4b18339c28e99e21d11041a55dd565cb918d" Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.345375 4884 scope.go:117] "RemoveContainer" containerID="1efcf5ffe8b307358ba20f5a5c0dd1d3beaac822091bb6ec447e6bc32003f950" Dec 02 02:46:07 crc kubenswrapper[4884]: E1202 02:46:07.345919 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1efcf5ffe8b307358ba20f5a5c0dd1d3beaac822091bb6ec447e6bc32003f950\": container with ID starting with 1efcf5ffe8b307358ba20f5a5c0dd1d3beaac822091bb6ec447e6bc32003f950 not found: ID does not exist" containerID="1efcf5ffe8b307358ba20f5a5c0dd1d3beaac822091bb6ec447e6bc32003f950" Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.345979 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1efcf5ffe8b307358ba20f5a5c0dd1d3beaac822091bb6ec447e6bc32003f950"} err="failed to get container status \"1efcf5ffe8b307358ba20f5a5c0dd1d3beaac822091bb6ec447e6bc32003f950\": rpc error: code = NotFound desc = could not find container \"1efcf5ffe8b307358ba20f5a5c0dd1d3beaac822091bb6ec447e6bc32003f950\": container with ID starting with 1efcf5ffe8b307358ba20f5a5c0dd1d3beaac822091bb6ec447e6bc32003f950 not found: ID does not exist" Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.346017 4884 scope.go:117] "RemoveContainer" containerID="e47bc953828954ab2f4f5e5b7dbb8deb44cfc3cf75740d0c9df409067e1c80a7" Dec 02 02:46:07 crc kubenswrapper[4884]: E1202 02:46:07.347241 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e47bc953828954ab2f4f5e5b7dbb8deb44cfc3cf75740d0c9df409067e1c80a7\": container with ID starting with e47bc953828954ab2f4f5e5b7dbb8deb44cfc3cf75740d0c9df409067e1c80a7 not found: ID does not exist" containerID="e47bc953828954ab2f4f5e5b7dbb8deb44cfc3cf75740d0c9df409067e1c80a7" Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.347290 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e47bc953828954ab2f4f5e5b7dbb8deb44cfc3cf75740d0c9df409067e1c80a7"} err="failed to get container status \"e47bc953828954ab2f4f5e5b7dbb8deb44cfc3cf75740d0c9df409067e1c80a7\": rpc error: code = NotFound desc = could not find container \"e47bc953828954ab2f4f5e5b7dbb8deb44cfc3cf75740d0c9df409067e1c80a7\": container with ID starting with e47bc953828954ab2f4f5e5b7dbb8deb44cfc3cf75740d0c9df409067e1c80a7 not found: ID does not exist" Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.347323 4884 scope.go:117] "RemoveContainer" containerID="df3fbc0e0ad84b1d7f090bec27bd4b18339c28e99e21d11041a55dd565cb918d" Dec 02 02:46:07 crc kubenswrapper[4884]: E1202 02:46:07.347848 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df3fbc0e0ad84b1d7f090bec27bd4b18339c28e99e21d11041a55dd565cb918d\": container with ID starting with df3fbc0e0ad84b1d7f090bec27bd4b18339c28e99e21d11041a55dd565cb918d not found: ID does not exist" containerID="df3fbc0e0ad84b1d7f090bec27bd4b18339c28e99e21d11041a55dd565cb918d" Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.347899 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df3fbc0e0ad84b1d7f090bec27bd4b18339c28e99e21d11041a55dd565cb918d"} err="failed to get container status \"df3fbc0e0ad84b1d7f090bec27bd4b18339c28e99e21d11041a55dd565cb918d\": rpc error: code = NotFound desc = could not find container \"df3fbc0e0ad84b1d7f090bec27bd4b18339c28e99e21d11041a55dd565cb918d\": container with ID starting with df3fbc0e0ad84b1d7f090bec27bd4b18339c28e99e21d11041a55dd565cb918d not found: ID does not exist" Dec 02 02:46:07 crc kubenswrapper[4884]: I1202 02:46:07.630772 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="791205ea-a0c8-476a-827d-5cbc1a24e774" path="/var/lib/kubelet/pods/791205ea-a0c8-476a-827d-5cbc1a24e774/volumes" Dec 02 02:47:46 crc kubenswrapper[4884]: I1202 02:47:46.971648 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:47:46 crc kubenswrapper[4884]: I1202 02:47:46.972958 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:48:16 crc kubenswrapper[4884]: I1202 02:48:16.972230 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:48:16 crc kubenswrapper[4884]: I1202 02:48:16.974625 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.438575 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-knnxc"] Dec 02 02:48:18 crc kubenswrapper[4884]: E1202 02:48:18.439313 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="791205ea-a0c8-476a-827d-5cbc1a24e774" containerName="extract-content" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.439336 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="791205ea-a0c8-476a-827d-5cbc1a24e774" containerName="extract-content" Dec 02 02:48:18 crc kubenswrapper[4884]: E1202 02:48:18.439373 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="791205ea-a0c8-476a-827d-5cbc1a24e774" containerName="extract-utilities" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.439388 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="791205ea-a0c8-476a-827d-5cbc1a24e774" containerName="extract-utilities" Dec 02 02:48:18 crc kubenswrapper[4884]: E1202 02:48:18.439408 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="791205ea-a0c8-476a-827d-5cbc1a24e774" containerName="registry-server" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.439421 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="791205ea-a0c8-476a-827d-5cbc1a24e774" containerName="registry-server" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.439782 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="791205ea-a0c8-476a-827d-5cbc1a24e774" containerName="registry-server" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.442382 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.470985 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-knnxc"] Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.492034 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/683b7fbb-6217-4e47-b7b9-6de55c3e812b-utilities\") pod \"redhat-marketplace-knnxc\" (UID: \"683b7fbb-6217-4e47-b7b9-6de55c3e812b\") " pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.492403 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4lct\" (UniqueName: \"kubernetes.io/projected/683b7fbb-6217-4e47-b7b9-6de55c3e812b-kube-api-access-l4lct\") pod \"redhat-marketplace-knnxc\" (UID: \"683b7fbb-6217-4e47-b7b9-6de55c3e812b\") " pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.492958 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/683b7fbb-6217-4e47-b7b9-6de55c3e812b-catalog-content\") pod \"redhat-marketplace-knnxc\" (UID: \"683b7fbb-6217-4e47-b7b9-6de55c3e812b\") " pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.594247 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4lct\" (UniqueName: \"kubernetes.io/projected/683b7fbb-6217-4e47-b7b9-6de55c3e812b-kube-api-access-l4lct\") pod \"redhat-marketplace-knnxc\" (UID: \"683b7fbb-6217-4e47-b7b9-6de55c3e812b\") " pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.594796 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/683b7fbb-6217-4e47-b7b9-6de55c3e812b-catalog-content\") pod \"redhat-marketplace-knnxc\" (UID: \"683b7fbb-6217-4e47-b7b9-6de55c3e812b\") " pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.594911 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/683b7fbb-6217-4e47-b7b9-6de55c3e812b-utilities\") pod \"redhat-marketplace-knnxc\" (UID: \"683b7fbb-6217-4e47-b7b9-6de55c3e812b\") " pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.595468 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/683b7fbb-6217-4e47-b7b9-6de55c3e812b-catalog-content\") pod \"redhat-marketplace-knnxc\" (UID: \"683b7fbb-6217-4e47-b7b9-6de55c3e812b\") " pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.595618 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/683b7fbb-6217-4e47-b7b9-6de55c3e812b-utilities\") pod \"redhat-marketplace-knnxc\" (UID: \"683b7fbb-6217-4e47-b7b9-6de55c3e812b\") " pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.636478 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sx4xc"] Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.638393 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.668067 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sx4xc"] Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.697860 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc43184-7204-4975-be42-f89e7f54c42d-utilities\") pod \"redhat-operators-sx4xc\" (UID: \"2dc43184-7204-4975-be42-f89e7f54c42d\") " pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.697906 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnrqk\" (UniqueName: \"kubernetes.io/projected/2dc43184-7204-4975-be42-f89e7f54c42d-kube-api-access-mnrqk\") pod \"redhat-operators-sx4xc\" (UID: \"2dc43184-7204-4975-be42-f89e7f54c42d\") " pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.698020 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc43184-7204-4975-be42-f89e7f54c42d-catalog-content\") pod \"redhat-operators-sx4xc\" (UID: \"2dc43184-7204-4975-be42-f89e7f54c42d\") " pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.749909 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="63a48d7f-b557-4f21-93d6-8a18e96eea16" containerName="galera" probeResult="failure" output="command timed out" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.750040 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="63a48d7f-b557-4f21-93d6-8a18e96eea16" containerName="galera" probeResult="failure" output="command timed out" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.752141 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4lct\" (UniqueName: \"kubernetes.io/projected/683b7fbb-6217-4e47-b7b9-6de55c3e812b-kube-api-access-l4lct\") pod \"redhat-marketplace-knnxc\" (UID: \"683b7fbb-6217-4e47-b7b9-6de55c3e812b\") " pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.800305 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc43184-7204-4975-be42-f89e7f54c42d-catalog-content\") pod \"redhat-operators-sx4xc\" (UID: \"2dc43184-7204-4975-be42-f89e7f54c42d\") " pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.800409 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc43184-7204-4975-be42-f89e7f54c42d-utilities\") pod \"redhat-operators-sx4xc\" (UID: \"2dc43184-7204-4975-be42-f89e7f54c42d\") " pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.800434 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnrqk\" (UniqueName: \"kubernetes.io/projected/2dc43184-7204-4975-be42-f89e7f54c42d-kube-api-access-mnrqk\") pod \"redhat-operators-sx4xc\" (UID: \"2dc43184-7204-4975-be42-f89e7f54c42d\") " pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.800760 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc43184-7204-4975-be42-f89e7f54c42d-catalog-content\") pod \"redhat-operators-sx4xc\" (UID: \"2dc43184-7204-4975-be42-f89e7f54c42d\") " pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.801003 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc43184-7204-4975-be42-f89e7f54c42d-utilities\") pod \"redhat-operators-sx4xc\" (UID: \"2dc43184-7204-4975-be42-f89e7f54c42d\") " pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.822023 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.822710 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnrqk\" (UniqueName: \"kubernetes.io/projected/2dc43184-7204-4975-be42-f89e7f54c42d-kube-api-access-mnrqk\") pod \"redhat-operators-sx4xc\" (UID: \"2dc43184-7204-4975-be42-f89e7f54c42d\") " pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:18 crc kubenswrapper[4884]: I1202 02:48:18.963640 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:19 crc kubenswrapper[4884]: I1202 02:48:19.389819 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-knnxc"] Dec 02 02:48:19 crc kubenswrapper[4884]: I1202 02:48:19.524593 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sx4xc"] Dec 02 02:48:19 crc kubenswrapper[4884]: W1202 02:48:19.529860 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2dc43184_7204_4975_be42_f89e7f54c42d.slice/crio-1935c61c056bb1c328367c94600b69ace612558ee4ab837070443b2d8488f839 WatchSource:0}: Error finding container 1935c61c056bb1c328367c94600b69ace612558ee4ab837070443b2d8488f839: Status 404 returned error can't find the container with id 1935c61c056bb1c328367c94600b69ace612558ee4ab837070443b2d8488f839 Dec 02 02:48:19 crc kubenswrapper[4884]: I1202 02:48:19.851309 4884 generic.go:334] "Generic (PLEG): container finished" podID="683b7fbb-6217-4e47-b7b9-6de55c3e812b" containerID="db9f99d9496839536490b213e1e8a269787e5ade8ae10beb44cd972c2c87c4a7" exitCode=0 Dec 02 02:48:19 crc kubenswrapper[4884]: I1202 02:48:19.851402 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-knnxc" event={"ID":"683b7fbb-6217-4e47-b7b9-6de55c3e812b","Type":"ContainerDied","Data":"db9f99d9496839536490b213e1e8a269787e5ade8ae10beb44cd972c2c87c4a7"} Dec 02 02:48:19 crc kubenswrapper[4884]: I1202 02:48:19.851428 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-knnxc" event={"ID":"683b7fbb-6217-4e47-b7b9-6de55c3e812b","Type":"ContainerStarted","Data":"e42edb9555be4cef55e2753f6275ab6d4c05cf47eeac28cfd5b8243ef97ddde7"} Dec 02 02:48:19 crc kubenswrapper[4884]: I1202 02:48:19.854311 4884 generic.go:334] "Generic (PLEG): container finished" podID="2dc43184-7204-4975-be42-f89e7f54c42d" containerID="c24d85313cd893d22fb930c48bfbba8d665d02e241f10736ec78f2d512fd0580" exitCode=0 Dec 02 02:48:19 crc kubenswrapper[4884]: I1202 02:48:19.854352 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx4xc" event={"ID":"2dc43184-7204-4975-be42-f89e7f54c42d","Type":"ContainerDied","Data":"c24d85313cd893d22fb930c48bfbba8d665d02e241f10736ec78f2d512fd0580"} Dec 02 02:48:19 crc kubenswrapper[4884]: I1202 02:48:19.854377 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx4xc" event={"ID":"2dc43184-7204-4975-be42-f89e7f54c42d","Type":"ContainerStarted","Data":"1935c61c056bb1c328367c94600b69ace612558ee4ab837070443b2d8488f839"} Dec 02 02:48:21 crc kubenswrapper[4884]: I1202 02:48:21.885096 4884 generic.go:334] "Generic (PLEG): container finished" podID="683b7fbb-6217-4e47-b7b9-6de55c3e812b" containerID="6309f436f433f57071e5b66db278cf555a31395096cd88e8b1e364b3067b4e88" exitCode=0 Dec 02 02:48:21 crc kubenswrapper[4884]: I1202 02:48:21.885229 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-knnxc" event={"ID":"683b7fbb-6217-4e47-b7b9-6de55c3e812b","Type":"ContainerDied","Data":"6309f436f433f57071e5b66db278cf555a31395096cd88e8b1e364b3067b4e88"} Dec 02 02:48:21 crc kubenswrapper[4884]: I1202 02:48:21.894019 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx4xc" event={"ID":"2dc43184-7204-4975-be42-f89e7f54c42d","Type":"ContainerStarted","Data":"1576ae1d8cc00cf3596a8805658daf1cc11fc07ec5b80b804b0f17797ff7e533"} Dec 02 02:48:23 crc kubenswrapper[4884]: I1202 02:48:23.918435 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-knnxc" event={"ID":"683b7fbb-6217-4e47-b7b9-6de55c3e812b","Type":"ContainerStarted","Data":"26479ea5ac94da6f4125b8c84433c4b3315d3e8f89ff94fd34bc2e946cbee9dd"} Dec 02 02:48:23 crc kubenswrapper[4884]: I1202 02:48:23.946989 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-knnxc" podStartSLOduration=3.269209907 podStartE2EDuration="5.946959413s" podCreationTimestamp="2025-12-02 02:48:18 +0000 UTC" firstStartedPulling="2025-12-02 02:48:19.853349889 +0000 UTC m=+4316.529186773" lastFinishedPulling="2025-12-02 02:48:22.531099375 +0000 UTC m=+4319.206936279" observedRunningTime="2025-12-02 02:48:23.938588998 +0000 UTC m=+4320.614425892" watchObservedRunningTime="2025-12-02 02:48:23.946959413 +0000 UTC m=+4320.622796317" Dec 02 02:48:24 crc kubenswrapper[4884]: I1202 02:48:24.932558 4884 generic.go:334] "Generic (PLEG): container finished" podID="2dc43184-7204-4975-be42-f89e7f54c42d" containerID="1576ae1d8cc00cf3596a8805658daf1cc11fc07ec5b80b804b0f17797ff7e533" exitCode=0 Dec 02 02:48:24 crc kubenswrapper[4884]: I1202 02:48:24.932628 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx4xc" event={"ID":"2dc43184-7204-4975-be42-f89e7f54c42d","Type":"ContainerDied","Data":"1576ae1d8cc00cf3596a8805658daf1cc11fc07ec5b80b804b0f17797ff7e533"} Dec 02 02:48:25 crc kubenswrapper[4884]: I1202 02:48:25.947619 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx4xc" event={"ID":"2dc43184-7204-4975-be42-f89e7f54c42d","Type":"ContainerStarted","Data":"0bb99136d24bc59c1f99f2bec773876c386e2dbc3e9acd571d769d26ccbe62fc"} Dec 02 02:48:25 crc kubenswrapper[4884]: I1202 02:48:25.988945 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sx4xc" podStartSLOduration=2.459596099 podStartE2EDuration="7.988923455s" podCreationTimestamp="2025-12-02 02:48:18 +0000 UTC" firstStartedPulling="2025-12-02 02:48:19.856489766 +0000 UTC m=+4316.532326650" lastFinishedPulling="2025-12-02 02:48:25.385817112 +0000 UTC m=+4322.061654006" observedRunningTime="2025-12-02 02:48:25.978210603 +0000 UTC m=+4322.654047497" watchObservedRunningTime="2025-12-02 02:48:25.988923455 +0000 UTC m=+4322.664760349" Dec 02 02:48:28 crc kubenswrapper[4884]: I1202 02:48:28.822888 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:28 crc kubenswrapper[4884]: I1202 02:48:28.823712 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:28 crc kubenswrapper[4884]: I1202 02:48:28.912971 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:28 crc kubenswrapper[4884]: I1202 02:48:28.964526 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:28 crc kubenswrapper[4884]: I1202 02:48:28.964606 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:29 crc kubenswrapper[4884]: I1202 02:48:29.075148 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:29 crc kubenswrapper[4884]: I1202 02:48:29.820775 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-knnxc"] Dec 02 02:48:30 crc kubenswrapper[4884]: I1202 02:48:30.031492 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-sx4xc" podUID="2dc43184-7204-4975-be42-f89e7f54c42d" containerName="registry-server" probeResult="failure" output=< Dec 02 02:48:30 crc kubenswrapper[4884]: timeout: failed to connect service ":50051" within 1s Dec 02 02:48:30 crc kubenswrapper[4884]: > Dec 02 02:48:31 crc kubenswrapper[4884]: I1202 02:48:31.014051 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-knnxc" podUID="683b7fbb-6217-4e47-b7b9-6de55c3e812b" containerName="registry-server" containerID="cri-o://26479ea5ac94da6f4125b8c84433c4b3315d3e8f89ff94fd34bc2e946cbee9dd" gracePeriod=2 Dec 02 02:48:31 crc kubenswrapper[4884]: I1202 02:48:31.588839 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:31 crc kubenswrapper[4884]: I1202 02:48:31.595331 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4lct\" (UniqueName: \"kubernetes.io/projected/683b7fbb-6217-4e47-b7b9-6de55c3e812b-kube-api-access-l4lct\") pod \"683b7fbb-6217-4e47-b7b9-6de55c3e812b\" (UID: \"683b7fbb-6217-4e47-b7b9-6de55c3e812b\") " Dec 02 02:48:31 crc kubenswrapper[4884]: I1202 02:48:31.595420 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/683b7fbb-6217-4e47-b7b9-6de55c3e812b-catalog-content\") pod \"683b7fbb-6217-4e47-b7b9-6de55c3e812b\" (UID: \"683b7fbb-6217-4e47-b7b9-6de55c3e812b\") " Dec 02 02:48:31 crc kubenswrapper[4884]: I1202 02:48:31.595590 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/683b7fbb-6217-4e47-b7b9-6de55c3e812b-utilities\") pod \"683b7fbb-6217-4e47-b7b9-6de55c3e812b\" (UID: \"683b7fbb-6217-4e47-b7b9-6de55c3e812b\") " Dec 02 02:48:31 crc kubenswrapper[4884]: I1202 02:48:31.597478 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/683b7fbb-6217-4e47-b7b9-6de55c3e812b-utilities" (OuterVolumeSpecName: "utilities") pod "683b7fbb-6217-4e47-b7b9-6de55c3e812b" (UID: "683b7fbb-6217-4e47-b7b9-6de55c3e812b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:48:31 crc kubenswrapper[4884]: I1202 02:48:31.606611 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/683b7fbb-6217-4e47-b7b9-6de55c3e812b-kube-api-access-l4lct" (OuterVolumeSpecName: "kube-api-access-l4lct") pod "683b7fbb-6217-4e47-b7b9-6de55c3e812b" (UID: "683b7fbb-6217-4e47-b7b9-6de55c3e812b"). InnerVolumeSpecName "kube-api-access-l4lct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:48:31 crc kubenswrapper[4884]: I1202 02:48:31.619025 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/683b7fbb-6217-4e47-b7b9-6de55c3e812b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "683b7fbb-6217-4e47-b7b9-6de55c3e812b" (UID: "683b7fbb-6217-4e47-b7b9-6de55c3e812b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:48:31 crc kubenswrapper[4884]: I1202 02:48:31.698008 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4lct\" (UniqueName: \"kubernetes.io/projected/683b7fbb-6217-4e47-b7b9-6de55c3e812b-kube-api-access-l4lct\") on node \"crc\" DevicePath \"\"" Dec 02 02:48:31 crc kubenswrapper[4884]: I1202 02:48:31.698036 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/683b7fbb-6217-4e47-b7b9-6de55c3e812b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:48:31 crc kubenswrapper[4884]: I1202 02:48:31.698046 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/683b7fbb-6217-4e47-b7b9-6de55c3e812b-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:48:31 crc kubenswrapper[4884]: E1202 02:48:31.740419 4884 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod683b7fbb_6217_4e47_b7b9_6de55c3e812b.slice\": RecentStats: unable to find data in memory cache]" Dec 02 02:48:32 crc kubenswrapper[4884]: I1202 02:48:32.025151 4884 generic.go:334] "Generic (PLEG): container finished" podID="683b7fbb-6217-4e47-b7b9-6de55c3e812b" containerID="26479ea5ac94da6f4125b8c84433c4b3315d3e8f89ff94fd34bc2e946cbee9dd" exitCode=0 Dec 02 02:48:32 crc kubenswrapper[4884]: I1202 02:48:32.025229 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-knnxc" Dec 02 02:48:32 crc kubenswrapper[4884]: I1202 02:48:32.025263 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-knnxc" event={"ID":"683b7fbb-6217-4e47-b7b9-6de55c3e812b","Type":"ContainerDied","Data":"26479ea5ac94da6f4125b8c84433c4b3315d3e8f89ff94fd34bc2e946cbee9dd"} Dec 02 02:48:32 crc kubenswrapper[4884]: I1202 02:48:32.025673 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-knnxc" event={"ID":"683b7fbb-6217-4e47-b7b9-6de55c3e812b","Type":"ContainerDied","Data":"e42edb9555be4cef55e2753f6275ab6d4c05cf47eeac28cfd5b8243ef97ddde7"} Dec 02 02:48:32 crc kubenswrapper[4884]: I1202 02:48:32.025705 4884 scope.go:117] "RemoveContainer" containerID="26479ea5ac94da6f4125b8c84433c4b3315d3e8f89ff94fd34bc2e946cbee9dd" Dec 02 02:48:32 crc kubenswrapper[4884]: I1202 02:48:32.050194 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-knnxc"] Dec 02 02:48:32 crc kubenswrapper[4884]: I1202 02:48:32.059906 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-knnxc"] Dec 02 02:48:32 crc kubenswrapper[4884]: I1202 02:48:32.060054 4884 scope.go:117] "RemoveContainer" containerID="6309f436f433f57071e5b66db278cf555a31395096cd88e8b1e364b3067b4e88" Dec 02 02:48:32 crc kubenswrapper[4884]: I1202 02:48:32.094361 4884 scope.go:117] "RemoveContainer" containerID="db9f99d9496839536490b213e1e8a269787e5ade8ae10beb44cd972c2c87c4a7" Dec 02 02:48:32 crc kubenswrapper[4884]: I1202 02:48:32.157397 4884 scope.go:117] "RemoveContainer" containerID="26479ea5ac94da6f4125b8c84433c4b3315d3e8f89ff94fd34bc2e946cbee9dd" Dec 02 02:48:32 crc kubenswrapper[4884]: E1202 02:48:32.158341 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26479ea5ac94da6f4125b8c84433c4b3315d3e8f89ff94fd34bc2e946cbee9dd\": container with ID starting with 26479ea5ac94da6f4125b8c84433c4b3315d3e8f89ff94fd34bc2e946cbee9dd not found: ID does not exist" containerID="26479ea5ac94da6f4125b8c84433c4b3315d3e8f89ff94fd34bc2e946cbee9dd" Dec 02 02:48:32 crc kubenswrapper[4884]: I1202 02:48:32.158387 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26479ea5ac94da6f4125b8c84433c4b3315d3e8f89ff94fd34bc2e946cbee9dd"} err="failed to get container status \"26479ea5ac94da6f4125b8c84433c4b3315d3e8f89ff94fd34bc2e946cbee9dd\": rpc error: code = NotFound desc = could not find container \"26479ea5ac94da6f4125b8c84433c4b3315d3e8f89ff94fd34bc2e946cbee9dd\": container with ID starting with 26479ea5ac94da6f4125b8c84433c4b3315d3e8f89ff94fd34bc2e946cbee9dd not found: ID does not exist" Dec 02 02:48:32 crc kubenswrapper[4884]: I1202 02:48:32.158478 4884 scope.go:117] "RemoveContainer" containerID="6309f436f433f57071e5b66db278cf555a31395096cd88e8b1e364b3067b4e88" Dec 02 02:48:32 crc kubenswrapper[4884]: E1202 02:48:32.159075 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6309f436f433f57071e5b66db278cf555a31395096cd88e8b1e364b3067b4e88\": container with ID starting with 6309f436f433f57071e5b66db278cf555a31395096cd88e8b1e364b3067b4e88 not found: ID does not exist" containerID="6309f436f433f57071e5b66db278cf555a31395096cd88e8b1e364b3067b4e88" Dec 02 02:48:32 crc kubenswrapper[4884]: I1202 02:48:32.159106 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6309f436f433f57071e5b66db278cf555a31395096cd88e8b1e364b3067b4e88"} err="failed to get container status \"6309f436f433f57071e5b66db278cf555a31395096cd88e8b1e364b3067b4e88\": rpc error: code = NotFound desc = could not find container \"6309f436f433f57071e5b66db278cf555a31395096cd88e8b1e364b3067b4e88\": container with ID starting with 6309f436f433f57071e5b66db278cf555a31395096cd88e8b1e364b3067b4e88 not found: ID does not exist" Dec 02 02:48:32 crc kubenswrapper[4884]: I1202 02:48:32.159127 4884 scope.go:117] "RemoveContainer" containerID="db9f99d9496839536490b213e1e8a269787e5ade8ae10beb44cd972c2c87c4a7" Dec 02 02:48:32 crc kubenswrapper[4884]: E1202 02:48:32.160646 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db9f99d9496839536490b213e1e8a269787e5ade8ae10beb44cd972c2c87c4a7\": container with ID starting with db9f99d9496839536490b213e1e8a269787e5ade8ae10beb44cd972c2c87c4a7 not found: ID does not exist" containerID="db9f99d9496839536490b213e1e8a269787e5ade8ae10beb44cd972c2c87c4a7" Dec 02 02:48:32 crc kubenswrapper[4884]: I1202 02:48:32.160692 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db9f99d9496839536490b213e1e8a269787e5ade8ae10beb44cd972c2c87c4a7"} err="failed to get container status \"db9f99d9496839536490b213e1e8a269787e5ade8ae10beb44cd972c2c87c4a7\": rpc error: code = NotFound desc = could not find container \"db9f99d9496839536490b213e1e8a269787e5ade8ae10beb44cd972c2c87c4a7\": container with ID starting with db9f99d9496839536490b213e1e8a269787e5ade8ae10beb44cd972c2c87c4a7 not found: ID does not exist" Dec 02 02:48:33 crc kubenswrapper[4884]: I1202 02:48:33.634801 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="683b7fbb-6217-4e47-b7b9-6de55c3e812b" path="/var/lib/kubelet/pods/683b7fbb-6217-4e47-b7b9-6de55c3e812b/volumes" Dec 02 02:48:39 crc kubenswrapper[4884]: I1202 02:48:39.032348 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:39 crc kubenswrapper[4884]: I1202 02:48:39.119429 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:39 crc kubenswrapper[4884]: I1202 02:48:39.279846 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sx4xc"] Dec 02 02:48:40 crc kubenswrapper[4884]: I1202 02:48:40.125059 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-sx4xc" podUID="2dc43184-7204-4975-be42-f89e7f54c42d" containerName="registry-server" containerID="cri-o://0bb99136d24bc59c1f99f2bec773876c386e2dbc3e9acd571d769d26ccbe62fc" gracePeriod=2 Dec 02 02:48:41 crc kubenswrapper[4884]: I1202 02:48:41.142233 4884 generic.go:334] "Generic (PLEG): container finished" podID="2dc43184-7204-4975-be42-f89e7f54c42d" containerID="0bb99136d24bc59c1f99f2bec773876c386e2dbc3e9acd571d769d26ccbe62fc" exitCode=0 Dec 02 02:48:41 crc kubenswrapper[4884]: I1202 02:48:41.142331 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx4xc" event={"ID":"2dc43184-7204-4975-be42-f89e7f54c42d","Type":"ContainerDied","Data":"0bb99136d24bc59c1f99f2bec773876c386e2dbc3e9acd571d769d26ccbe62fc"} Dec 02 02:48:41 crc kubenswrapper[4884]: I1202 02:48:41.142564 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sx4xc" event={"ID":"2dc43184-7204-4975-be42-f89e7f54c42d","Type":"ContainerDied","Data":"1935c61c056bb1c328367c94600b69ace612558ee4ab837070443b2d8488f839"} Dec 02 02:48:41 crc kubenswrapper[4884]: I1202 02:48:41.142589 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1935c61c056bb1c328367c94600b69ace612558ee4ab837070443b2d8488f839" Dec 02 02:48:41 crc kubenswrapper[4884]: I1202 02:48:41.196140 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:41 crc kubenswrapper[4884]: I1202 02:48:41.230046 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc43184-7204-4975-be42-f89e7f54c42d-utilities\") pod \"2dc43184-7204-4975-be42-f89e7f54c42d\" (UID: \"2dc43184-7204-4975-be42-f89e7f54c42d\") " Dec 02 02:48:41 crc kubenswrapper[4884]: I1202 02:48:41.230138 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc43184-7204-4975-be42-f89e7f54c42d-catalog-content\") pod \"2dc43184-7204-4975-be42-f89e7f54c42d\" (UID: \"2dc43184-7204-4975-be42-f89e7f54c42d\") " Dec 02 02:48:41 crc kubenswrapper[4884]: I1202 02:48:41.230185 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrqk\" (UniqueName: \"kubernetes.io/projected/2dc43184-7204-4975-be42-f89e7f54c42d-kube-api-access-mnrqk\") pod \"2dc43184-7204-4975-be42-f89e7f54c42d\" (UID: \"2dc43184-7204-4975-be42-f89e7f54c42d\") " Dec 02 02:48:41 crc kubenswrapper[4884]: I1202 02:48:41.231471 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dc43184-7204-4975-be42-f89e7f54c42d-utilities" (OuterVolumeSpecName: "utilities") pod "2dc43184-7204-4975-be42-f89e7f54c42d" (UID: "2dc43184-7204-4975-be42-f89e7f54c42d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:48:41 crc kubenswrapper[4884]: I1202 02:48:41.232363 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2dc43184-7204-4975-be42-f89e7f54c42d-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:48:41 crc kubenswrapper[4884]: I1202 02:48:41.248826 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dc43184-7204-4975-be42-f89e7f54c42d-kube-api-access-mnrqk" (OuterVolumeSpecName: "kube-api-access-mnrqk") pod "2dc43184-7204-4975-be42-f89e7f54c42d" (UID: "2dc43184-7204-4975-be42-f89e7f54c42d"). InnerVolumeSpecName "kube-api-access-mnrqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:48:41 crc kubenswrapper[4884]: I1202 02:48:41.334966 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrqk\" (UniqueName: \"kubernetes.io/projected/2dc43184-7204-4975-be42-f89e7f54c42d-kube-api-access-mnrqk\") on node \"crc\" DevicePath \"\"" Dec 02 02:48:41 crc kubenswrapper[4884]: I1202 02:48:41.385072 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2dc43184-7204-4975-be42-f89e7f54c42d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2dc43184-7204-4975-be42-f89e7f54c42d" (UID: "2dc43184-7204-4975-be42-f89e7f54c42d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:48:41 crc kubenswrapper[4884]: I1202 02:48:41.435970 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2dc43184-7204-4975-be42-f89e7f54c42d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:48:42 crc kubenswrapper[4884]: I1202 02:48:42.158305 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sx4xc" Dec 02 02:48:42 crc kubenswrapper[4884]: I1202 02:48:42.201955 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-sx4xc"] Dec 02 02:48:42 crc kubenswrapper[4884]: I1202 02:48:42.214218 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-sx4xc"] Dec 02 02:48:43 crc kubenswrapper[4884]: I1202 02:48:43.659473 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dc43184-7204-4975-be42-f89e7f54c42d" path="/var/lib/kubelet/pods/2dc43184-7204-4975-be42-f89e7f54c42d/volumes" Dec 02 02:48:46 crc kubenswrapper[4884]: I1202 02:48:46.971965 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:48:46 crc kubenswrapper[4884]: I1202 02:48:46.972702 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:48:46 crc kubenswrapper[4884]: I1202 02:48:46.972851 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 02:48:46 crc kubenswrapper[4884]: I1202 02:48:46.973794 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"579c75b60c975cce5b8ce8cb9ee141c65e4945b53bab8840e41aec5e8d052b1c"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 02:48:46 crc kubenswrapper[4884]: I1202 02:48:46.973893 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://579c75b60c975cce5b8ce8cb9ee141c65e4945b53bab8840e41aec5e8d052b1c" gracePeriod=600 Dec 02 02:48:47 crc kubenswrapper[4884]: I1202 02:48:47.224270 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="579c75b60c975cce5b8ce8cb9ee141c65e4945b53bab8840e41aec5e8d052b1c" exitCode=0 Dec 02 02:48:47 crc kubenswrapper[4884]: I1202 02:48:47.224511 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"579c75b60c975cce5b8ce8cb9ee141c65e4945b53bab8840e41aec5e8d052b1c"} Dec 02 02:48:47 crc kubenswrapper[4884]: I1202 02:48:47.224790 4884 scope.go:117] "RemoveContainer" containerID="e28f66e0f9c838419175e234efec54f34e5402d45f460fa7a1f293c616a6a3aa" Dec 02 02:48:48 crc kubenswrapper[4884]: I1202 02:48:48.257241 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b"} Dec 02 02:49:37 crc kubenswrapper[4884]: I1202 02:49:37.889601 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-28zpb"] Dec 02 02:49:37 crc kubenswrapper[4884]: E1202 02:49:37.891021 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dc43184-7204-4975-be42-f89e7f54c42d" containerName="extract-utilities" Dec 02 02:49:37 crc kubenswrapper[4884]: I1202 02:49:37.891046 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dc43184-7204-4975-be42-f89e7f54c42d" containerName="extract-utilities" Dec 02 02:49:37 crc kubenswrapper[4884]: E1202 02:49:37.891071 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dc43184-7204-4975-be42-f89e7f54c42d" containerName="extract-content" Dec 02 02:49:37 crc kubenswrapper[4884]: I1202 02:49:37.891083 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dc43184-7204-4975-be42-f89e7f54c42d" containerName="extract-content" Dec 02 02:49:37 crc kubenswrapper[4884]: E1202 02:49:37.891104 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="683b7fbb-6217-4e47-b7b9-6de55c3e812b" containerName="registry-server" Dec 02 02:49:37 crc kubenswrapper[4884]: I1202 02:49:37.891117 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="683b7fbb-6217-4e47-b7b9-6de55c3e812b" containerName="registry-server" Dec 02 02:49:37 crc kubenswrapper[4884]: E1202 02:49:37.891173 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dc43184-7204-4975-be42-f89e7f54c42d" containerName="registry-server" Dec 02 02:49:37 crc kubenswrapper[4884]: I1202 02:49:37.891185 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dc43184-7204-4975-be42-f89e7f54c42d" containerName="registry-server" Dec 02 02:49:37 crc kubenswrapper[4884]: E1202 02:49:37.891212 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="683b7fbb-6217-4e47-b7b9-6de55c3e812b" containerName="extract-utilities" Dec 02 02:49:37 crc kubenswrapper[4884]: I1202 02:49:37.891226 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="683b7fbb-6217-4e47-b7b9-6de55c3e812b" containerName="extract-utilities" Dec 02 02:49:37 crc kubenswrapper[4884]: E1202 02:49:37.891241 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="683b7fbb-6217-4e47-b7b9-6de55c3e812b" containerName="extract-content" Dec 02 02:49:37 crc kubenswrapper[4884]: I1202 02:49:37.891253 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="683b7fbb-6217-4e47-b7b9-6de55c3e812b" containerName="extract-content" Dec 02 02:49:37 crc kubenswrapper[4884]: I1202 02:49:37.891665 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dc43184-7204-4975-be42-f89e7f54c42d" containerName="registry-server" Dec 02 02:49:37 crc kubenswrapper[4884]: I1202 02:49:37.891709 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="683b7fbb-6217-4e47-b7b9-6de55c3e812b" containerName="registry-server" Dec 02 02:49:37 crc kubenswrapper[4884]: I1202 02:49:37.894677 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:37 crc kubenswrapper[4884]: I1202 02:49:37.930820 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-28zpb"] Dec 02 02:49:37 crc kubenswrapper[4884]: I1202 02:49:37.983416 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cvj2\" (UniqueName: \"kubernetes.io/projected/363112ca-b211-4395-bdc0-63387b3fe60d-kube-api-access-6cvj2\") pod \"certified-operators-28zpb\" (UID: \"363112ca-b211-4395-bdc0-63387b3fe60d\") " pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:37 crc kubenswrapper[4884]: I1202 02:49:37.983513 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/363112ca-b211-4395-bdc0-63387b3fe60d-catalog-content\") pod \"certified-operators-28zpb\" (UID: \"363112ca-b211-4395-bdc0-63387b3fe60d\") " pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:37 crc kubenswrapper[4884]: I1202 02:49:37.983726 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/363112ca-b211-4395-bdc0-63387b3fe60d-utilities\") pod \"certified-operators-28zpb\" (UID: \"363112ca-b211-4395-bdc0-63387b3fe60d\") " pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:38 crc kubenswrapper[4884]: I1202 02:49:38.085491 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cvj2\" (UniqueName: \"kubernetes.io/projected/363112ca-b211-4395-bdc0-63387b3fe60d-kube-api-access-6cvj2\") pod \"certified-operators-28zpb\" (UID: \"363112ca-b211-4395-bdc0-63387b3fe60d\") " pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:38 crc kubenswrapper[4884]: I1202 02:49:38.085583 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/363112ca-b211-4395-bdc0-63387b3fe60d-catalog-content\") pod \"certified-operators-28zpb\" (UID: \"363112ca-b211-4395-bdc0-63387b3fe60d\") " pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:38 crc kubenswrapper[4884]: I1202 02:49:38.085827 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/363112ca-b211-4395-bdc0-63387b3fe60d-utilities\") pod \"certified-operators-28zpb\" (UID: \"363112ca-b211-4395-bdc0-63387b3fe60d\") " pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:38 crc kubenswrapper[4884]: I1202 02:49:38.086169 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/363112ca-b211-4395-bdc0-63387b3fe60d-catalog-content\") pod \"certified-operators-28zpb\" (UID: \"363112ca-b211-4395-bdc0-63387b3fe60d\") " pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:38 crc kubenswrapper[4884]: I1202 02:49:38.086237 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/363112ca-b211-4395-bdc0-63387b3fe60d-utilities\") pod \"certified-operators-28zpb\" (UID: \"363112ca-b211-4395-bdc0-63387b3fe60d\") " pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:38 crc kubenswrapper[4884]: I1202 02:49:38.107248 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cvj2\" (UniqueName: \"kubernetes.io/projected/363112ca-b211-4395-bdc0-63387b3fe60d-kube-api-access-6cvj2\") pod \"certified-operators-28zpb\" (UID: \"363112ca-b211-4395-bdc0-63387b3fe60d\") " pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:38 crc kubenswrapper[4884]: I1202 02:49:38.231696 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:38 crc kubenswrapper[4884]: I1202 02:49:38.777654 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-28zpb"] Dec 02 02:49:38 crc kubenswrapper[4884]: I1202 02:49:38.902436 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28zpb" event={"ID":"363112ca-b211-4395-bdc0-63387b3fe60d","Type":"ContainerStarted","Data":"6ef4c6b1e5b042ac1d8ef2dcf0596aba313d96593e1cbf5ce3b66e72022f380c"} Dec 02 02:49:39 crc kubenswrapper[4884]: I1202 02:49:39.919632 4884 generic.go:334] "Generic (PLEG): container finished" podID="363112ca-b211-4395-bdc0-63387b3fe60d" containerID="7ce2a00b3c3a5569e78e43b7b7e321cac56f8fbfcf4dc98f5206e09aed940b49" exitCode=0 Dec 02 02:49:39 crc kubenswrapper[4884]: I1202 02:49:39.919730 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28zpb" event={"ID":"363112ca-b211-4395-bdc0-63387b3fe60d","Type":"ContainerDied","Data":"7ce2a00b3c3a5569e78e43b7b7e321cac56f8fbfcf4dc98f5206e09aed940b49"} Dec 02 02:49:41 crc kubenswrapper[4884]: I1202 02:49:41.946657 4884 generic.go:334] "Generic (PLEG): container finished" podID="363112ca-b211-4395-bdc0-63387b3fe60d" containerID="182bd082ee38129f060b16c4669c773a473d48d9afe76f9391e5c0d9f4f2a123" exitCode=0 Dec 02 02:49:41 crc kubenswrapper[4884]: I1202 02:49:41.946789 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28zpb" event={"ID":"363112ca-b211-4395-bdc0-63387b3fe60d","Type":"ContainerDied","Data":"182bd082ee38129f060b16c4669c773a473d48d9afe76f9391e5c0d9f4f2a123"} Dec 02 02:49:42 crc kubenswrapper[4884]: I1202 02:49:42.971446 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28zpb" event={"ID":"363112ca-b211-4395-bdc0-63387b3fe60d","Type":"ContainerStarted","Data":"989731523b5988b4929e9c5352844a746fcef44aee9b44fb1ec32262b9f79058"} Dec 02 02:49:43 crc kubenswrapper[4884]: I1202 02:49:43.000685 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-28zpb" podStartSLOduration=3.421575794 podStartE2EDuration="6.000658376s" podCreationTimestamp="2025-12-02 02:49:37 +0000 UTC" firstStartedPulling="2025-12-02 02:49:39.924086758 +0000 UTC m=+4396.599923652" lastFinishedPulling="2025-12-02 02:49:42.50316934 +0000 UTC m=+4399.179006234" observedRunningTime="2025-12-02 02:49:42.997472338 +0000 UTC m=+4399.673309232" watchObservedRunningTime="2025-12-02 02:49:43.000658376 +0000 UTC m=+4399.676495270" Dec 02 02:49:48 crc kubenswrapper[4884]: I1202 02:49:48.232475 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:48 crc kubenswrapper[4884]: I1202 02:49:48.233231 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:49 crc kubenswrapper[4884]: I1202 02:49:49.152035 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:49 crc kubenswrapper[4884]: I1202 02:49:49.279384 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:49 crc kubenswrapper[4884]: I1202 02:49:49.399949 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-28zpb"] Dec 02 02:49:51 crc kubenswrapper[4884]: I1202 02:49:51.063569 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-28zpb" podUID="363112ca-b211-4395-bdc0-63387b3fe60d" containerName="registry-server" containerID="cri-o://989731523b5988b4929e9c5352844a746fcef44aee9b44fb1ec32262b9f79058" gracePeriod=2 Dec 02 02:49:51 crc kubenswrapper[4884]: I1202 02:49:51.640452 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:51 crc kubenswrapper[4884]: I1202 02:49:51.740627 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/363112ca-b211-4395-bdc0-63387b3fe60d-utilities\") pod \"363112ca-b211-4395-bdc0-63387b3fe60d\" (UID: \"363112ca-b211-4395-bdc0-63387b3fe60d\") " Dec 02 02:49:51 crc kubenswrapper[4884]: I1202 02:49:51.740814 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/363112ca-b211-4395-bdc0-63387b3fe60d-catalog-content\") pod \"363112ca-b211-4395-bdc0-63387b3fe60d\" (UID: \"363112ca-b211-4395-bdc0-63387b3fe60d\") " Dec 02 02:49:51 crc kubenswrapper[4884]: I1202 02:49:51.740881 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cvj2\" (UniqueName: \"kubernetes.io/projected/363112ca-b211-4395-bdc0-63387b3fe60d-kube-api-access-6cvj2\") pod \"363112ca-b211-4395-bdc0-63387b3fe60d\" (UID: \"363112ca-b211-4395-bdc0-63387b3fe60d\") " Dec 02 02:49:51 crc kubenswrapper[4884]: I1202 02:49:51.741722 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/363112ca-b211-4395-bdc0-63387b3fe60d-utilities" (OuterVolumeSpecName: "utilities") pod "363112ca-b211-4395-bdc0-63387b3fe60d" (UID: "363112ca-b211-4395-bdc0-63387b3fe60d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:49:51 crc kubenswrapper[4884]: I1202 02:49:51.746142 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/363112ca-b211-4395-bdc0-63387b3fe60d-kube-api-access-6cvj2" (OuterVolumeSpecName: "kube-api-access-6cvj2") pod "363112ca-b211-4395-bdc0-63387b3fe60d" (UID: "363112ca-b211-4395-bdc0-63387b3fe60d"). InnerVolumeSpecName "kube-api-access-6cvj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:49:51 crc kubenswrapper[4884]: I1202 02:49:51.843375 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cvj2\" (UniqueName: \"kubernetes.io/projected/363112ca-b211-4395-bdc0-63387b3fe60d-kube-api-access-6cvj2\") on node \"crc\" DevicePath \"\"" Dec 02 02:49:51 crc kubenswrapper[4884]: I1202 02:49:51.843415 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/363112ca-b211-4395-bdc0-63387b3fe60d-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.077282 4884 generic.go:334] "Generic (PLEG): container finished" podID="363112ca-b211-4395-bdc0-63387b3fe60d" containerID="989731523b5988b4929e9c5352844a746fcef44aee9b44fb1ec32262b9f79058" exitCode=0 Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.077355 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28zpb" event={"ID":"363112ca-b211-4395-bdc0-63387b3fe60d","Type":"ContainerDied","Data":"989731523b5988b4929e9c5352844a746fcef44aee9b44fb1ec32262b9f79058"} Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.077424 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-28zpb" event={"ID":"363112ca-b211-4395-bdc0-63387b3fe60d","Type":"ContainerDied","Data":"6ef4c6b1e5b042ac1d8ef2dcf0596aba313d96593e1cbf5ce3b66e72022f380c"} Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.077371 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-28zpb" Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.077454 4884 scope.go:117] "RemoveContainer" containerID="989731523b5988b4929e9c5352844a746fcef44aee9b44fb1ec32262b9f79058" Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.099126 4884 scope.go:117] "RemoveContainer" containerID="182bd082ee38129f060b16c4669c773a473d48d9afe76f9391e5c0d9f4f2a123" Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.164161 4884 scope.go:117] "RemoveContainer" containerID="7ce2a00b3c3a5569e78e43b7b7e321cac56f8fbfcf4dc98f5206e09aed940b49" Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.200899 4884 scope.go:117] "RemoveContainer" containerID="989731523b5988b4929e9c5352844a746fcef44aee9b44fb1ec32262b9f79058" Dec 02 02:49:52 crc kubenswrapper[4884]: E1202 02:49:52.202364 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"989731523b5988b4929e9c5352844a746fcef44aee9b44fb1ec32262b9f79058\": container with ID starting with 989731523b5988b4929e9c5352844a746fcef44aee9b44fb1ec32262b9f79058 not found: ID does not exist" containerID="989731523b5988b4929e9c5352844a746fcef44aee9b44fb1ec32262b9f79058" Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.202440 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"989731523b5988b4929e9c5352844a746fcef44aee9b44fb1ec32262b9f79058"} err="failed to get container status \"989731523b5988b4929e9c5352844a746fcef44aee9b44fb1ec32262b9f79058\": rpc error: code = NotFound desc = could not find container \"989731523b5988b4929e9c5352844a746fcef44aee9b44fb1ec32262b9f79058\": container with ID starting with 989731523b5988b4929e9c5352844a746fcef44aee9b44fb1ec32262b9f79058 not found: ID does not exist" Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.202485 4884 scope.go:117] "RemoveContainer" containerID="182bd082ee38129f060b16c4669c773a473d48d9afe76f9391e5c0d9f4f2a123" Dec 02 02:49:52 crc kubenswrapper[4884]: E1202 02:49:52.202939 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"182bd082ee38129f060b16c4669c773a473d48d9afe76f9391e5c0d9f4f2a123\": container with ID starting with 182bd082ee38129f060b16c4669c773a473d48d9afe76f9391e5c0d9f4f2a123 not found: ID does not exist" containerID="182bd082ee38129f060b16c4669c773a473d48d9afe76f9391e5c0d9f4f2a123" Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.203009 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"182bd082ee38129f060b16c4669c773a473d48d9afe76f9391e5c0d9f4f2a123"} err="failed to get container status \"182bd082ee38129f060b16c4669c773a473d48d9afe76f9391e5c0d9f4f2a123\": rpc error: code = NotFound desc = could not find container \"182bd082ee38129f060b16c4669c773a473d48d9afe76f9391e5c0d9f4f2a123\": container with ID starting with 182bd082ee38129f060b16c4669c773a473d48d9afe76f9391e5c0d9f4f2a123 not found: ID does not exist" Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.203058 4884 scope.go:117] "RemoveContainer" containerID="7ce2a00b3c3a5569e78e43b7b7e321cac56f8fbfcf4dc98f5206e09aed940b49" Dec 02 02:49:52 crc kubenswrapper[4884]: E1202 02:49:52.203845 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ce2a00b3c3a5569e78e43b7b7e321cac56f8fbfcf4dc98f5206e09aed940b49\": container with ID starting with 7ce2a00b3c3a5569e78e43b7b7e321cac56f8fbfcf4dc98f5206e09aed940b49 not found: ID does not exist" containerID="7ce2a00b3c3a5569e78e43b7b7e321cac56f8fbfcf4dc98f5206e09aed940b49" Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.203890 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ce2a00b3c3a5569e78e43b7b7e321cac56f8fbfcf4dc98f5206e09aed940b49"} err="failed to get container status \"7ce2a00b3c3a5569e78e43b7b7e321cac56f8fbfcf4dc98f5206e09aed940b49\": rpc error: code = NotFound desc = could not find container \"7ce2a00b3c3a5569e78e43b7b7e321cac56f8fbfcf4dc98f5206e09aed940b49\": container with ID starting with 7ce2a00b3c3a5569e78e43b7b7e321cac56f8fbfcf4dc98f5206e09aed940b49 not found: ID does not exist" Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.216716 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/363112ca-b211-4395-bdc0-63387b3fe60d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "363112ca-b211-4395-bdc0-63387b3fe60d" (UID: "363112ca-b211-4395-bdc0-63387b3fe60d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.251071 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/363112ca-b211-4395-bdc0-63387b3fe60d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.450807 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-28zpb"] Dec 02 02:49:52 crc kubenswrapper[4884]: I1202 02:49:52.462259 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-28zpb"] Dec 02 02:49:53 crc kubenswrapper[4884]: I1202 02:49:53.633989 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="363112ca-b211-4395-bdc0-63387b3fe60d" path="/var/lib/kubelet/pods/363112ca-b211-4395-bdc0-63387b3fe60d/volumes" Dec 02 02:50:48 crc kubenswrapper[4884]: I1202 02:50:48.665868 4884 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-558bcd5597-pglnv" podUID="c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 02 02:51:16 crc kubenswrapper[4884]: I1202 02:51:16.972002 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:51:16 crc kubenswrapper[4884]: I1202 02:51:16.972550 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:51:46 crc kubenswrapper[4884]: I1202 02:51:46.971961 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:51:46 crc kubenswrapper[4884]: I1202 02:51:46.972582 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:52:16 crc kubenswrapper[4884]: I1202 02:52:16.972013 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:52:16 crc kubenswrapper[4884]: I1202 02:52:16.972889 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 02:52:16 crc kubenswrapper[4884]: I1202 02:52:16.973004 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 02:52:16 crc kubenswrapper[4884]: I1202 02:52:16.974395 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 02:52:16 crc kubenswrapper[4884]: I1202 02:52:16.974533 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" gracePeriod=600 Dec 02 02:52:17 crc kubenswrapper[4884]: E1202 02:52:17.108658 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:52:17 crc kubenswrapper[4884]: I1202 02:52:17.926122 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" exitCode=0 Dec 02 02:52:17 crc kubenswrapper[4884]: I1202 02:52:17.926220 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b"} Dec 02 02:52:17 crc kubenswrapper[4884]: I1202 02:52:17.927138 4884 scope.go:117] "RemoveContainer" containerID="579c75b60c975cce5b8ce8cb9ee141c65e4945b53bab8840e41aec5e8d052b1c" Dec 02 02:52:17 crc kubenswrapper[4884]: I1202 02:52:17.928196 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:52:17 crc kubenswrapper[4884]: E1202 02:52:17.928811 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:52:29 crc kubenswrapper[4884]: I1202 02:52:29.614837 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:52:29 crc kubenswrapper[4884]: E1202 02:52:29.615938 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:52:40 crc kubenswrapper[4884]: I1202 02:52:40.616191 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:52:40 crc kubenswrapper[4884]: E1202 02:52:40.616848 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:52:54 crc kubenswrapper[4884]: I1202 02:52:54.615115 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:52:54 crc kubenswrapper[4884]: E1202 02:52:54.618505 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:53:05 crc kubenswrapper[4884]: I1202 02:53:05.615824 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:53:05 crc kubenswrapper[4884]: E1202 02:53:05.616821 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:53:18 crc kubenswrapper[4884]: I1202 02:53:18.615876 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:53:18 crc kubenswrapper[4884]: E1202 02:53:18.616845 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:53:30 crc kubenswrapper[4884]: I1202 02:53:30.614194 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:53:30 crc kubenswrapper[4884]: E1202 02:53:30.616297 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:53:45 crc kubenswrapper[4884]: I1202 02:53:45.614967 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:53:45 crc kubenswrapper[4884]: E1202 02:53:45.616153 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:53:56 crc kubenswrapper[4884]: I1202 02:53:56.615370 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:53:56 crc kubenswrapper[4884]: E1202 02:53:56.617526 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:54:07 crc kubenswrapper[4884]: I1202 02:54:07.614265 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:54:07 crc kubenswrapper[4884]: E1202 02:54:07.614969 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:54:22 crc kubenswrapper[4884]: I1202 02:54:22.618916 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:54:22 crc kubenswrapper[4884]: E1202 02:54:22.620423 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:54:34 crc kubenswrapper[4884]: I1202 02:54:34.615080 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:54:34 crc kubenswrapper[4884]: E1202 02:54:34.616139 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:54:46 crc kubenswrapper[4884]: I1202 02:54:46.615215 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:54:46 crc kubenswrapper[4884]: E1202 02:54:46.616483 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:54:47 crc kubenswrapper[4884]: I1202 02:54:47.306149 4884 scope.go:117] "RemoveContainer" containerID="c24d85313cd893d22fb930c48bfbba8d665d02e241f10736ec78f2d512fd0580" Dec 02 02:54:47 crc kubenswrapper[4884]: I1202 02:54:47.331969 4884 scope.go:117] "RemoveContainer" containerID="1576ae1d8cc00cf3596a8805658daf1cc11fc07ec5b80b804b0f17797ff7e533" Dec 02 02:54:47 crc kubenswrapper[4884]: I1202 02:54:47.401898 4884 scope.go:117] "RemoveContainer" containerID="0bb99136d24bc59c1f99f2bec773876c386e2dbc3e9acd571d769d26ccbe62fc" Dec 02 02:55:01 crc kubenswrapper[4884]: I1202 02:55:01.614976 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:55:01 crc kubenswrapper[4884]: E1202 02:55:01.616078 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:55:13 crc kubenswrapper[4884]: I1202 02:55:13.630203 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:55:13 crc kubenswrapper[4884]: E1202 02:55:13.631607 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:55:28 crc kubenswrapper[4884]: I1202 02:55:28.614868 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:55:28 crc kubenswrapper[4884]: E1202 02:55:28.615728 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:55:43 crc kubenswrapper[4884]: I1202 02:55:43.626486 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:55:43 crc kubenswrapper[4884]: E1202 02:55:43.627466 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:55:56 crc kubenswrapper[4884]: I1202 02:55:56.615220 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:55:56 crc kubenswrapper[4884]: E1202 02:55:56.616006 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:56:10 crc kubenswrapper[4884]: I1202 02:56:10.615102 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:56:10 crc kubenswrapper[4884]: E1202 02:56:10.616370 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:56:21 crc kubenswrapper[4884]: I1202 02:56:21.614946 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:56:21 crc kubenswrapper[4884]: E1202 02:56:21.615789 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:56:35 crc kubenswrapper[4884]: I1202 02:56:35.616043 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:56:35 crc kubenswrapper[4884]: E1202 02:56:35.616981 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:56:50 crc kubenswrapper[4884]: I1202 02:56:50.614218 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:56:50 crc kubenswrapper[4884]: E1202 02:56:50.617462 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:56:57 crc kubenswrapper[4884]: I1202 02:56:57.942686 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rstmz"] Dec 02 02:56:57 crc kubenswrapper[4884]: E1202 02:56:57.944131 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363112ca-b211-4395-bdc0-63387b3fe60d" containerName="extract-utilities" Dec 02 02:56:57 crc kubenswrapper[4884]: I1202 02:56:57.944154 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="363112ca-b211-4395-bdc0-63387b3fe60d" containerName="extract-utilities" Dec 02 02:56:57 crc kubenswrapper[4884]: E1202 02:56:57.944179 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363112ca-b211-4395-bdc0-63387b3fe60d" containerName="extract-content" Dec 02 02:56:57 crc kubenswrapper[4884]: I1202 02:56:57.944192 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="363112ca-b211-4395-bdc0-63387b3fe60d" containerName="extract-content" Dec 02 02:56:57 crc kubenswrapper[4884]: E1202 02:56:57.944226 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363112ca-b211-4395-bdc0-63387b3fe60d" containerName="registry-server" Dec 02 02:56:57 crc kubenswrapper[4884]: I1202 02:56:57.944239 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="363112ca-b211-4395-bdc0-63387b3fe60d" containerName="registry-server" Dec 02 02:56:57 crc kubenswrapper[4884]: I1202 02:56:57.944595 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="363112ca-b211-4395-bdc0-63387b3fe60d" containerName="registry-server" Dec 02 02:56:57 crc kubenswrapper[4884]: I1202 02:56:57.947166 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:56:57 crc kubenswrapper[4884]: I1202 02:56:57.957810 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rstmz"] Dec 02 02:56:58 crc kubenswrapper[4884]: I1202 02:56:58.070513 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-catalog-content\") pod \"community-operators-rstmz\" (UID: \"ace0fa8b-9483-4e75-813d-ef98a33bb4f2\") " pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:56:58 crc kubenswrapper[4884]: I1202 02:56:58.070693 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjcdh\" (UniqueName: \"kubernetes.io/projected/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-kube-api-access-mjcdh\") pod \"community-operators-rstmz\" (UID: \"ace0fa8b-9483-4e75-813d-ef98a33bb4f2\") " pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:56:58 crc kubenswrapper[4884]: I1202 02:56:58.070868 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-utilities\") pod \"community-operators-rstmz\" (UID: \"ace0fa8b-9483-4e75-813d-ef98a33bb4f2\") " pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:56:58 crc kubenswrapper[4884]: I1202 02:56:58.173568 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-utilities\") pod \"community-operators-rstmz\" (UID: \"ace0fa8b-9483-4e75-813d-ef98a33bb4f2\") " pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:56:58 crc kubenswrapper[4884]: I1202 02:56:58.173737 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-catalog-content\") pod \"community-operators-rstmz\" (UID: \"ace0fa8b-9483-4e75-813d-ef98a33bb4f2\") " pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:56:58 crc kubenswrapper[4884]: I1202 02:56:58.173872 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjcdh\" (UniqueName: \"kubernetes.io/projected/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-kube-api-access-mjcdh\") pod \"community-operators-rstmz\" (UID: \"ace0fa8b-9483-4e75-813d-ef98a33bb4f2\") " pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:56:58 crc kubenswrapper[4884]: I1202 02:56:58.174275 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-utilities\") pod \"community-operators-rstmz\" (UID: \"ace0fa8b-9483-4e75-813d-ef98a33bb4f2\") " pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:56:58 crc kubenswrapper[4884]: I1202 02:56:58.174300 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-catalog-content\") pod \"community-operators-rstmz\" (UID: \"ace0fa8b-9483-4e75-813d-ef98a33bb4f2\") " pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:56:58 crc kubenswrapper[4884]: I1202 02:56:58.197264 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjcdh\" (UniqueName: \"kubernetes.io/projected/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-kube-api-access-mjcdh\") pod \"community-operators-rstmz\" (UID: \"ace0fa8b-9483-4e75-813d-ef98a33bb4f2\") " pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:56:58 crc kubenswrapper[4884]: I1202 02:56:58.275077 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:56:58 crc kubenswrapper[4884]: I1202 02:56:58.830966 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rstmz"] Dec 02 02:56:59 crc kubenswrapper[4884]: I1202 02:56:59.810658 4884 generic.go:334] "Generic (PLEG): container finished" podID="ace0fa8b-9483-4e75-813d-ef98a33bb4f2" containerID="40b26c5b2515884efa936cd288404a071ef80b4922c4e270e2015dc953eee419" exitCode=0 Dec 02 02:56:59 crc kubenswrapper[4884]: I1202 02:56:59.810807 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rstmz" event={"ID":"ace0fa8b-9483-4e75-813d-ef98a33bb4f2","Type":"ContainerDied","Data":"40b26c5b2515884efa936cd288404a071ef80b4922c4e270e2015dc953eee419"} Dec 02 02:56:59 crc kubenswrapper[4884]: I1202 02:56:59.811078 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rstmz" event={"ID":"ace0fa8b-9483-4e75-813d-ef98a33bb4f2","Type":"ContainerStarted","Data":"3e2ab5fc67eef2f5a4ebc96b67fb4e5f95ef8f1151d2753b9f8301c92a0c848b"} Dec 02 02:56:59 crc kubenswrapper[4884]: I1202 02:56:59.814118 4884 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 02:57:01 crc kubenswrapper[4884]: I1202 02:57:01.614958 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:57:01 crc kubenswrapper[4884]: E1202 02:57:01.615877 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:57:01 crc kubenswrapper[4884]: I1202 02:57:01.840440 4884 generic.go:334] "Generic (PLEG): container finished" podID="ace0fa8b-9483-4e75-813d-ef98a33bb4f2" containerID="61e684562588dc20858aa286881190b9f2c9cf92cc4b56ff1044f4a49316e1cc" exitCode=0 Dec 02 02:57:01 crc kubenswrapper[4884]: I1202 02:57:01.840504 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rstmz" event={"ID":"ace0fa8b-9483-4e75-813d-ef98a33bb4f2","Type":"ContainerDied","Data":"61e684562588dc20858aa286881190b9f2c9cf92cc4b56ff1044f4a49316e1cc"} Dec 02 02:57:02 crc kubenswrapper[4884]: I1202 02:57:02.854068 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rstmz" event={"ID":"ace0fa8b-9483-4e75-813d-ef98a33bb4f2","Type":"ContainerStarted","Data":"c4afa1e7fe0be5d9f57cba87ce0a101a5a3f2b0ced5d818b3a1ccf93910d7d04"} Dec 02 02:57:02 crc kubenswrapper[4884]: I1202 02:57:02.881484 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rstmz" podStartSLOduration=3.410967679 podStartE2EDuration="5.881468123s" podCreationTimestamp="2025-12-02 02:56:57 +0000 UTC" firstStartedPulling="2025-12-02 02:56:59.813702119 +0000 UTC m=+4836.489539033" lastFinishedPulling="2025-12-02 02:57:02.284202553 +0000 UTC m=+4838.960039477" observedRunningTime="2025-12-02 02:57:02.874304267 +0000 UTC m=+4839.550141181" watchObservedRunningTime="2025-12-02 02:57:02.881468123 +0000 UTC m=+4839.557305007" Dec 02 02:57:08 crc kubenswrapper[4884]: I1202 02:57:08.276111 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:57:08 crc kubenswrapper[4884]: I1202 02:57:08.276835 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:57:08 crc kubenswrapper[4884]: I1202 02:57:08.360266 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:57:09 crc kubenswrapper[4884]: I1202 02:57:09.306545 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:57:09 crc kubenswrapper[4884]: I1202 02:57:09.389326 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rstmz"] Dec 02 02:57:10 crc kubenswrapper[4884]: I1202 02:57:10.954221 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rstmz" podUID="ace0fa8b-9483-4e75-813d-ef98a33bb4f2" containerName="registry-server" containerID="cri-o://c4afa1e7fe0be5d9f57cba87ce0a101a5a3f2b0ced5d818b3a1ccf93910d7d04" gracePeriod=2 Dec 02 02:57:11 crc kubenswrapper[4884]: I1202 02:57:11.618143 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:57:11 crc kubenswrapper[4884]: I1202 02:57:11.716981 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-utilities\") pod \"ace0fa8b-9483-4e75-813d-ef98a33bb4f2\" (UID: \"ace0fa8b-9483-4e75-813d-ef98a33bb4f2\") " Dec 02 02:57:11 crc kubenswrapper[4884]: I1202 02:57:11.717061 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjcdh\" (UniqueName: \"kubernetes.io/projected/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-kube-api-access-mjcdh\") pod \"ace0fa8b-9483-4e75-813d-ef98a33bb4f2\" (UID: \"ace0fa8b-9483-4e75-813d-ef98a33bb4f2\") " Dec 02 02:57:11 crc kubenswrapper[4884]: I1202 02:57:11.717359 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-catalog-content\") pod \"ace0fa8b-9483-4e75-813d-ef98a33bb4f2\" (UID: \"ace0fa8b-9483-4e75-813d-ef98a33bb4f2\") " Dec 02 02:57:11 crc kubenswrapper[4884]: I1202 02:57:11.720651 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-utilities" (OuterVolumeSpecName: "utilities") pod "ace0fa8b-9483-4e75-813d-ef98a33bb4f2" (UID: "ace0fa8b-9483-4e75-813d-ef98a33bb4f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:57:11 crc kubenswrapper[4884]: I1202 02:57:11.744712 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-kube-api-access-mjcdh" (OuterVolumeSpecName: "kube-api-access-mjcdh") pod "ace0fa8b-9483-4e75-813d-ef98a33bb4f2" (UID: "ace0fa8b-9483-4e75-813d-ef98a33bb4f2"). InnerVolumeSpecName "kube-api-access-mjcdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:57:11 crc kubenswrapper[4884]: I1202 02:57:11.793653 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ace0fa8b-9483-4e75-813d-ef98a33bb4f2" (UID: "ace0fa8b-9483-4e75-813d-ef98a33bb4f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:57:11 crc kubenswrapper[4884]: I1202 02:57:11.819838 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:57:11 crc kubenswrapper[4884]: I1202 02:57:11.819864 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:57:11 crc kubenswrapper[4884]: I1202 02:57:11.819873 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjcdh\" (UniqueName: \"kubernetes.io/projected/ace0fa8b-9483-4e75-813d-ef98a33bb4f2-kube-api-access-mjcdh\") on node \"crc\" DevicePath \"\"" Dec 02 02:57:11 crc kubenswrapper[4884]: I1202 02:57:11.974634 4884 generic.go:334] "Generic (PLEG): container finished" podID="ace0fa8b-9483-4e75-813d-ef98a33bb4f2" containerID="c4afa1e7fe0be5d9f57cba87ce0a101a5a3f2b0ced5d818b3a1ccf93910d7d04" exitCode=0 Dec 02 02:57:11 crc kubenswrapper[4884]: I1202 02:57:11.974862 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rstmz" event={"ID":"ace0fa8b-9483-4e75-813d-ef98a33bb4f2","Type":"ContainerDied","Data":"c4afa1e7fe0be5d9f57cba87ce0a101a5a3f2b0ced5d818b3a1ccf93910d7d04"} Dec 02 02:57:11 crc kubenswrapper[4884]: I1202 02:57:11.975074 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rstmz" event={"ID":"ace0fa8b-9483-4e75-813d-ef98a33bb4f2","Type":"ContainerDied","Data":"3e2ab5fc67eef2f5a4ebc96b67fb4e5f95ef8f1151d2753b9f8301c92a0c848b"} Dec 02 02:57:11 crc kubenswrapper[4884]: I1202 02:57:11.974962 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rstmz" Dec 02 02:57:11 crc kubenswrapper[4884]: I1202 02:57:11.975110 4884 scope.go:117] "RemoveContainer" containerID="c4afa1e7fe0be5d9f57cba87ce0a101a5a3f2b0ced5d818b3a1ccf93910d7d04" Dec 02 02:57:12 crc kubenswrapper[4884]: I1202 02:57:12.019537 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rstmz"] Dec 02 02:57:12 crc kubenswrapper[4884]: I1202 02:57:12.023928 4884 scope.go:117] "RemoveContainer" containerID="61e684562588dc20858aa286881190b9f2c9cf92cc4b56ff1044f4a49316e1cc" Dec 02 02:57:12 crc kubenswrapper[4884]: I1202 02:57:12.034885 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rstmz"] Dec 02 02:57:12 crc kubenswrapper[4884]: I1202 02:57:12.062807 4884 scope.go:117] "RemoveContainer" containerID="40b26c5b2515884efa936cd288404a071ef80b4922c4e270e2015dc953eee419" Dec 02 02:57:12 crc kubenswrapper[4884]: I1202 02:57:12.112281 4884 scope.go:117] "RemoveContainer" containerID="c4afa1e7fe0be5d9f57cba87ce0a101a5a3f2b0ced5d818b3a1ccf93910d7d04" Dec 02 02:57:12 crc kubenswrapper[4884]: E1202 02:57:12.112795 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4afa1e7fe0be5d9f57cba87ce0a101a5a3f2b0ced5d818b3a1ccf93910d7d04\": container with ID starting with c4afa1e7fe0be5d9f57cba87ce0a101a5a3f2b0ced5d818b3a1ccf93910d7d04 not found: ID does not exist" containerID="c4afa1e7fe0be5d9f57cba87ce0a101a5a3f2b0ced5d818b3a1ccf93910d7d04" Dec 02 02:57:12 crc kubenswrapper[4884]: I1202 02:57:12.112833 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4afa1e7fe0be5d9f57cba87ce0a101a5a3f2b0ced5d818b3a1ccf93910d7d04"} err="failed to get container status \"c4afa1e7fe0be5d9f57cba87ce0a101a5a3f2b0ced5d818b3a1ccf93910d7d04\": rpc error: code = NotFound desc = could not find container \"c4afa1e7fe0be5d9f57cba87ce0a101a5a3f2b0ced5d818b3a1ccf93910d7d04\": container with ID starting with c4afa1e7fe0be5d9f57cba87ce0a101a5a3f2b0ced5d818b3a1ccf93910d7d04 not found: ID does not exist" Dec 02 02:57:12 crc kubenswrapper[4884]: I1202 02:57:12.112855 4884 scope.go:117] "RemoveContainer" containerID="61e684562588dc20858aa286881190b9f2c9cf92cc4b56ff1044f4a49316e1cc" Dec 02 02:57:12 crc kubenswrapper[4884]: E1202 02:57:12.113534 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61e684562588dc20858aa286881190b9f2c9cf92cc4b56ff1044f4a49316e1cc\": container with ID starting with 61e684562588dc20858aa286881190b9f2c9cf92cc4b56ff1044f4a49316e1cc not found: ID does not exist" containerID="61e684562588dc20858aa286881190b9f2c9cf92cc4b56ff1044f4a49316e1cc" Dec 02 02:57:12 crc kubenswrapper[4884]: I1202 02:57:12.113626 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61e684562588dc20858aa286881190b9f2c9cf92cc4b56ff1044f4a49316e1cc"} err="failed to get container status \"61e684562588dc20858aa286881190b9f2c9cf92cc4b56ff1044f4a49316e1cc\": rpc error: code = NotFound desc = could not find container \"61e684562588dc20858aa286881190b9f2c9cf92cc4b56ff1044f4a49316e1cc\": container with ID starting with 61e684562588dc20858aa286881190b9f2c9cf92cc4b56ff1044f4a49316e1cc not found: ID does not exist" Dec 02 02:57:12 crc kubenswrapper[4884]: I1202 02:57:12.113680 4884 scope.go:117] "RemoveContainer" containerID="40b26c5b2515884efa936cd288404a071ef80b4922c4e270e2015dc953eee419" Dec 02 02:57:12 crc kubenswrapper[4884]: E1202 02:57:12.114184 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40b26c5b2515884efa936cd288404a071ef80b4922c4e270e2015dc953eee419\": container with ID starting with 40b26c5b2515884efa936cd288404a071ef80b4922c4e270e2015dc953eee419 not found: ID does not exist" containerID="40b26c5b2515884efa936cd288404a071ef80b4922c4e270e2015dc953eee419" Dec 02 02:57:12 crc kubenswrapper[4884]: I1202 02:57:12.114209 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40b26c5b2515884efa936cd288404a071ef80b4922c4e270e2015dc953eee419"} err="failed to get container status \"40b26c5b2515884efa936cd288404a071ef80b4922c4e270e2015dc953eee419\": rpc error: code = NotFound desc = could not find container \"40b26c5b2515884efa936cd288404a071ef80b4922c4e270e2015dc953eee419\": container with ID starting with 40b26c5b2515884efa936cd288404a071ef80b4922c4e270e2015dc953eee419 not found: ID does not exist" Dec 02 02:57:12 crc kubenswrapper[4884]: I1202 02:57:12.614360 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:57:12 crc kubenswrapper[4884]: E1202 02:57:12.614822 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 02:57:13 crc kubenswrapper[4884]: I1202 02:57:13.637056 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ace0fa8b-9483-4e75-813d-ef98a33bb4f2" path="/var/lib/kubelet/pods/ace0fa8b-9483-4e75-813d-ef98a33bb4f2/volumes" Dec 02 02:57:27 crc kubenswrapper[4884]: I1202 02:57:27.615490 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 02:57:28 crc kubenswrapper[4884]: I1202 02:57:28.173051 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"1d849541abc4fdea729a5ec715545cbcd4dffd71b8ffc714c87d3321a01e2b9a"} Dec 02 02:59:01 crc kubenswrapper[4884]: I1202 02:59:01.907131 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wvc99"] Dec 02 02:59:01 crc kubenswrapper[4884]: E1202 02:59:01.908347 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ace0fa8b-9483-4e75-813d-ef98a33bb4f2" containerName="registry-server" Dec 02 02:59:01 crc kubenswrapper[4884]: I1202 02:59:01.908368 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ace0fa8b-9483-4e75-813d-ef98a33bb4f2" containerName="registry-server" Dec 02 02:59:01 crc kubenswrapper[4884]: E1202 02:59:01.908398 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ace0fa8b-9483-4e75-813d-ef98a33bb4f2" containerName="extract-utilities" Dec 02 02:59:01 crc kubenswrapper[4884]: I1202 02:59:01.908411 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ace0fa8b-9483-4e75-813d-ef98a33bb4f2" containerName="extract-utilities" Dec 02 02:59:01 crc kubenswrapper[4884]: E1202 02:59:01.908461 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ace0fa8b-9483-4e75-813d-ef98a33bb4f2" containerName="extract-content" Dec 02 02:59:01 crc kubenswrapper[4884]: I1202 02:59:01.908476 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ace0fa8b-9483-4e75-813d-ef98a33bb4f2" containerName="extract-content" Dec 02 02:59:01 crc kubenswrapper[4884]: I1202 02:59:01.908911 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="ace0fa8b-9483-4e75-813d-ef98a33bb4f2" containerName="registry-server" Dec 02 02:59:01 crc kubenswrapper[4884]: I1202 02:59:01.911808 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:01 crc kubenswrapper[4884]: I1202 02:59:01.936700 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvc99"] Dec 02 02:59:01 crc kubenswrapper[4884]: I1202 02:59:01.974833 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l85j9\" (UniqueName: \"kubernetes.io/projected/74c85300-3f81-4a20-8dd5-a8a95816e132-kube-api-access-l85j9\") pod \"redhat-marketplace-wvc99\" (UID: \"74c85300-3f81-4a20-8dd5-a8a95816e132\") " pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:01 crc kubenswrapper[4884]: I1202 02:59:01.975300 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74c85300-3f81-4a20-8dd5-a8a95816e132-utilities\") pod \"redhat-marketplace-wvc99\" (UID: \"74c85300-3f81-4a20-8dd5-a8a95816e132\") " pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:01 crc kubenswrapper[4884]: I1202 02:59:01.976173 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74c85300-3f81-4a20-8dd5-a8a95816e132-catalog-content\") pod \"redhat-marketplace-wvc99\" (UID: \"74c85300-3f81-4a20-8dd5-a8a95816e132\") " pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.078647 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74c85300-3f81-4a20-8dd5-a8a95816e132-utilities\") pod \"redhat-marketplace-wvc99\" (UID: \"74c85300-3f81-4a20-8dd5-a8a95816e132\") " pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.078994 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74c85300-3f81-4a20-8dd5-a8a95816e132-catalog-content\") pod \"redhat-marketplace-wvc99\" (UID: \"74c85300-3f81-4a20-8dd5-a8a95816e132\") " pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.079141 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l85j9\" (UniqueName: \"kubernetes.io/projected/74c85300-3f81-4a20-8dd5-a8a95816e132-kube-api-access-l85j9\") pod \"redhat-marketplace-wvc99\" (UID: \"74c85300-3f81-4a20-8dd5-a8a95816e132\") " pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.079320 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74c85300-3f81-4a20-8dd5-a8a95816e132-utilities\") pod \"redhat-marketplace-wvc99\" (UID: \"74c85300-3f81-4a20-8dd5-a8a95816e132\") " pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.079657 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74c85300-3f81-4a20-8dd5-a8a95816e132-catalog-content\") pod \"redhat-marketplace-wvc99\" (UID: \"74c85300-3f81-4a20-8dd5-a8a95816e132\") " pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.101334 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l85j9\" (UniqueName: \"kubernetes.io/projected/74c85300-3f81-4a20-8dd5-a8a95816e132-kube-api-access-l85j9\") pod \"redhat-marketplace-wvc99\" (UID: \"74c85300-3f81-4a20-8dd5-a8a95816e132\") " pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.109380 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x7bx8"] Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.111995 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.139534 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7bx8"] Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.181368 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-catalog-content\") pod \"redhat-operators-x7bx8\" (UID: \"5ab5af5b-5b74-44af-a0bc-698aa76a47d4\") " pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.181509 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-utilities\") pod \"redhat-operators-x7bx8\" (UID: \"5ab5af5b-5b74-44af-a0bc-698aa76a47d4\") " pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.181547 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l4th\" (UniqueName: \"kubernetes.io/projected/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-kube-api-access-8l4th\") pod \"redhat-operators-x7bx8\" (UID: \"5ab5af5b-5b74-44af-a0bc-698aa76a47d4\") " pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.236793 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.283812 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-utilities\") pod \"redhat-operators-x7bx8\" (UID: \"5ab5af5b-5b74-44af-a0bc-698aa76a47d4\") " pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.283874 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l4th\" (UniqueName: \"kubernetes.io/projected/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-kube-api-access-8l4th\") pod \"redhat-operators-x7bx8\" (UID: \"5ab5af5b-5b74-44af-a0bc-698aa76a47d4\") " pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.283953 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-catalog-content\") pod \"redhat-operators-x7bx8\" (UID: \"5ab5af5b-5b74-44af-a0bc-698aa76a47d4\") " pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.284676 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-catalog-content\") pod \"redhat-operators-x7bx8\" (UID: \"5ab5af5b-5b74-44af-a0bc-698aa76a47d4\") " pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.285691 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-utilities\") pod \"redhat-operators-x7bx8\" (UID: \"5ab5af5b-5b74-44af-a0bc-698aa76a47d4\") " pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.302930 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l4th\" (UniqueName: \"kubernetes.io/projected/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-kube-api-access-8l4th\") pod \"redhat-operators-x7bx8\" (UID: \"5ab5af5b-5b74-44af-a0bc-698aa76a47d4\") " pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.486976 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.802413 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvc99"] Dec 02 02:59:02 crc kubenswrapper[4884]: I1202 02:59:02.987314 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x7bx8"] Dec 02 02:59:03 crc kubenswrapper[4884]: I1202 02:59:03.287140 4884 generic.go:334] "Generic (PLEG): container finished" podID="74c85300-3f81-4a20-8dd5-a8a95816e132" containerID="8639902764e9d3e836acbb0db3ade4689d085e3b79410839612d18b4f6d4545b" exitCode=0 Dec 02 02:59:03 crc kubenswrapper[4884]: I1202 02:59:03.287257 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvc99" event={"ID":"74c85300-3f81-4a20-8dd5-a8a95816e132","Type":"ContainerDied","Data":"8639902764e9d3e836acbb0db3ade4689d085e3b79410839612d18b4f6d4545b"} Dec 02 02:59:03 crc kubenswrapper[4884]: I1202 02:59:03.287725 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvc99" event={"ID":"74c85300-3f81-4a20-8dd5-a8a95816e132","Type":"ContainerStarted","Data":"6b0c7190bbbd6479de8ea2fee4f0095d34cd47c565c960203623ba0458591be4"} Dec 02 02:59:03 crc kubenswrapper[4884]: I1202 02:59:03.289492 4884 generic.go:334] "Generic (PLEG): container finished" podID="5ab5af5b-5b74-44af-a0bc-698aa76a47d4" containerID="0c7ceed92bb9427b4d57fbc5cc29d437e8bc2e50adeb81fd6817632d940f8158" exitCode=0 Dec 02 02:59:03 crc kubenswrapper[4884]: I1202 02:59:03.289556 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7bx8" event={"ID":"5ab5af5b-5b74-44af-a0bc-698aa76a47d4","Type":"ContainerDied","Data":"0c7ceed92bb9427b4d57fbc5cc29d437e8bc2e50adeb81fd6817632d940f8158"} Dec 02 02:59:03 crc kubenswrapper[4884]: I1202 02:59:03.289649 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7bx8" event={"ID":"5ab5af5b-5b74-44af-a0bc-698aa76a47d4","Type":"ContainerStarted","Data":"216051586d6ff0fd905e787ea529e39e29692827c2b85f6cc423df276f6d85bf"} Dec 02 02:59:05 crc kubenswrapper[4884]: I1202 02:59:05.316350 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvc99" event={"ID":"74c85300-3f81-4a20-8dd5-a8a95816e132","Type":"ContainerStarted","Data":"d95d50ae17d040d4b82dee640807bcf3ca7dd48c4e64145823055b229a350472"} Dec 02 02:59:05 crc kubenswrapper[4884]: I1202 02:59:05.320578 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7bx8" event={"ID":"5ab5af5b-5b74-44af-a0bc-698aa76a47d4","Type":"ContainerStarted","Data":"6b39bd32afeabd6ccef761a109b5ffd91572d43ef1bf68ce77dee227ff72841c"} Dec 02 02:59:06 crc kubenswrapper[4884]: I1202 02:59:06.335670 4884 generic.go:334] "Generic (PLEG): container finished" podID="74c85300-3f81-4a20-8dd5-a8a95816e132" containerID="d95d50ae17d040d4b82dee640807bcf3ca7dd48c4e64145823055b229a350472" exitCode=0 Dec 02 02:59:06 crc kubenswrapper[4884]: I1202 02:59:06.335775 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvc99" event={"ID":"74c85300-3f81-4a20-8dd5-a8a95816e132","Type":"ContainerDied","Data":"d95d50ae17d040d4b82dee640807bcf3ca7dd48c4e64145823055b229a350472"} Dec 02 02:59:07 crc kubenswrapper[4884]: I1202 02:59:07.346500 4884 generic.go:334] "Generic (PLEG): container finished" podID="5ab5af5b-5b74-44af-a0bc-698aa76a47d4" containerID="6b39bd32afeabd6ccef761a109b5ffd91572d43ef1bf68ce77dee227ff72841c" exitCode=0 Dec 02 02:59:07 crc kubenswrapper[4884]: I1202 02:59:07.346541 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7bx8" event={"ID":"5ab5af5b-5b74-44af-a0bc-698aa76a47d4","Type":"ContainerDied","Data":"6b39bd32afeabd6ccef761a109b5ffd91572d43ef1bf68ce77dee227ff72841c"} Dec 02 02:59:08 crc kubenswrapper[4884]: I1202 02:59:08.357359 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvc99" event={"ID":"74c85300-3f81-4a20-8dd5-a8a95816e132","Type":"ContainerStarted","Data":"f4ee66d68589e96d094f771d3b67099e70e9ff554dc6a2116eafb761325faa95"} Dec 02 02:59:08 crc kubenswrapper[4884]: I1202 02:59:08.377174 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wvc99" podStartSLOduration=3.162666942 podStartE2EDuration="7.377154394s" podCreationTimestamp="2025-12-02 02:59:01 +0000 UTC" firstStartedPulling="2025-12-02 02:59:03.288650808 +0000 UTC m=+4959.964487692" lastFinishedPulling="2025-12-02 02:59:07.50313826 +0000 UTC m=+4964.178975144" observedRunningTime="2025-12-02 02:59:08.373872694 +0000 UTC m=+4965.049709598" watchObservedRunningTime="2025-12-02 02:59:08.377154394 +0000 UTC m=+4965.052991298" Dec 02 02:59:09 crc kubenswrapper[4884]: I1202 02:59:09.369414 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7bx8" event={"ID":"5ab5af5b-5b74-44af-a0bc-698aa76a47d4","Type":"ContainerStarted","Data":"82a3a2434948e7a4caa31f6fdc5ad3427ce0230f9a9cf4e8a947d5e0bc953ffa"} Dec 02 02:59:09 crc kubenswrapper[4884]: I1202 02:59:09.393565 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x7bx8" podStartSLOduration=2.794248762 podStartE2EDuration="7.393540843s" podCreationTimestamp="2025-12-02 02:59:02 +0000 UTC" firstStartedPulling="2025-12-02 02:59:03.292277297 +0000 UTC m=+4959.968114181" lastFinishedPulling="2025-12-02 02:59:07.891569368 +0000 UTC m=+4964.567406262" observedRunningTime="2025-12-02 02:59:09.388486109 +0000 UTC m=+4966.064323043" watchObservedRunningTime="2025-12-02 02:59:09.393540843 +0000 UTC m=+4966.069377767" Dec 02 02:59:12 crc kubenswrapper[4884]: I1202 02:59:12.236926 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:12 crc kubenswrapper[4884]: I1202 02:59:12.237398 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:12 crc kubenswrapper[4884]: I1202 02:59:12.363215 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:12 crc kubenswrapper[4884]: I1202 02:59:12.462209 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:12 crc kubenswrapper[4884]: I1202 02:59:12.491824 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:12 crc kubenswrapper[4884]: I1202 02:59:12.493033 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:12 crc kubenswrapper[4884]: I1202 02:59:12.894592 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvc99"] Dec 02 02:59:13 crc kubenswrapper[4884]: I1202 02:59:13.555668 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x7bx8" podUID="5ab5af5b-5b74-44af-a0bc-698aa76a47d4" containerName="registry-server" probeResult="failure" output=< Dec 02 02:59:13 crc kubenswrapper[4884]: timeout: failed to connect service ":50051" within 1s Dec 02 02:59:13 crc kubenswrapper[4884]: > Dec 02 02:59:14 crc kubenswrapper[4884]: I1202 02:59:14.417560 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wvc99" podUID="74c85300-3f81-4a20-8dd5-a8a95816e132" containerName="registry-server" containerID="cri-o://f4ee66d68589e96d094f771d3b67099e70e9ff554dc6a2116eafb761325faa95" gracePeriod=2 Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.048305 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.152982 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74c85300-3f81-4a20-8dd5-a8a95816e132-utilities\") pod \"74c85300-3f81-4a20-8dd5-a8a95816e132\" (UID: \"74c85300-3f81-4a20-8dd5-a8a95816e132\") " Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.153039 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l85j9\" (UniqueName: \"kubernetes.io/projected/74c85300-3f81-4a20-8dd5-a8a95816e132-kube-api-access-l85j9\") pod \"74c85300-3f81-4a20-8dd5-a8a95816e132\" (UID: \"74c85300-3f81-4a20-8dd5-a8a95816e132\") " Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.153076 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74c85300-3f81-4a20-8dd5-a8a95816e132-catalog-content\") pod \"74c85300-3f81-4a20-8dd5-a8a95816e132\" (UID: \"74c85300-3f81-4a20-8dd5-a8a95816e132\") " Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.164775 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74c85300-3f81-4a20-8dd5-a8a95816e132-kube-api-access-l85j9" (OuterVolumeSpecName: "kube-api-access-l85j9") pod "74c85300-3f81-4a20-8dd5-a8a95816e132" (UID: "74c85300-3f81-4a20-8dd5-a8a95816e132"). InnerVolumeSpecName "kube-api-access-l85j9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.168240 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74c85300-3f81-4a20-8dd5-a8a95816e132-utilities" (OuterVolumeSpecName: "utilities") pod "74c85300-3f81-4a20-8dd5-a8a95816e132" (UID: "74c85300-3f81-4a20-8dd5-a8a95816e132"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.186272 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74c85300-3f81-4a20-8dd5-a8a95816e132-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74c85300-3f81-4a20-8dd5-a8a95816e132" (UID: "74c85300-3f81-4a20-8dd5-a8a95816e132"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.255230 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74c85300-3f81-4a20-8dd5-a8a95816e132-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.255275 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l85j9\" (UniqueName: \"kubernetes.io/projected/74c85300-3f81-4a20-8dd5-a8a95816e132-kube-api-access-l85j9\") on node \"crc\" DevicePath \"\"" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.255286 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74c85300-3f81-4a20-8dd5-a8a95816e132-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.429348 4884 generic.go:334] "Generic (PLEG): container finished" podID="74c85300-3f81-4a20-8dd5-a8a95816e132" containerID="f4ee66d68589e96d094f771d3b67099e70e9ff554dc6a2116eafb761325faa95" exitCode=0 Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.429457 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wvc99" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.429432 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvc99" event={"ID":"74c85300-3f81-4a20-8dd5-a8a95816e132","Type":"ContainerDied","Data":"f4ee66d68589e96d094f771d3b67099e70e9ff554dc6a2116eafb761325faa95"} Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.429873 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wvc99" event={"ID":"74c85300-3f81-4a20-8dd5-a8a95816e132","Type":"ContainerDied","Data":"6b0c7190bbbd6479de8ea2fee4f0095d34cd47c565c960203623ba0458591be4"} Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.429915 4884 scope.go:117] "RemoveContainer" containerID="f4ee66d68589e96d094f771d3b67099e70e9ff554dc6a2116eafb761325faa95" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.460113 4884 scope.go:117] "RemoveContainer" containerID="d95d50ae17d040d4b82dee640807bcf3ca7dd48c4e64145823055b229a350472" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.465895 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvc99"] Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.473056 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wvc99"] Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.501972 4884 scope.go:117] "RemoveContainer" containerID="8639902764e9d3e836acbb0db3ade4689d085e3b79410839612d18b4f6d4545b" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.535927 4884 scope.go:117] "RemoveContainer" containerID="f4ee66d68589e96d094f771d3b67099e70e9ff554dc6a2116eafb761325faa95" Dec 02 02:59:15 crc kubenswrapper[4884]: E1202 02:59:15.536442 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4ee66d68589e96d094f771d3b67099e70e9ff554dc6a2116eafb761325faa95\": container with ID starting with f4ee66d68589e96d094f771d3b67099e70e9ff554dc6a2116eafb761325faa95 not found: ID does not exist" containerID="f4ee66d68589e96d094f771d3b67099e70e9ff554dc6a2116eafb761325faa95" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.536531 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4ee66d68589e96d094f771d3b67099e70e9ff554dc6a2116eafb761325faa95"} err="failed to get container status \"f4ee66d68589e96d094f771d3b67099e70e9ff554dc6a2116eafb761325faa95\": rpc error: code = NotFound desc = could not find container \"f4ee66d68589e96d094f771d3b67099e70e9ff554dc6a2116eafb761325faa95\": container with ID starting with f4ee66d68589e96d094f771d3b67099e70e9ff554dc6a2116eafb761325faa95 not found: ID does not exist" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.536603 4884 scope.go:117] "RemoveContainer" containerID="d95d50ae17d040d4b82dee640807bcf3ca7dd48c4e64145823055b229a350472" Dec 02 02:59:15 crc kubenswrapper[4884]: E1202 02:59:15.537002 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d95d50ae17d040d4b82dee640807bcf3ca7dd48c4e64145823055b229a350472\": container with ID starting with d95d50ae17d040d4b82dee640807bcf3ca7dd48c4e64145823055b229a350472 not found: ID does not exist" containerID="d95d50ae17d040d4b82dee640807bcf3ca7dd48c4e64145823055b229a350472" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.537023 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d95d50ae17d040d4b82dee640807bcf3ca7dd48c4e64145823055b229a350472"} err="failed to get container status \"d95d50ae17d040d4b82dee640807bcf3ca7dd48c4e64145823055b229a350472\": rpc error: code = NotFound desc = could not find container \"d95d50ae17d040d4b82dee640807bcf3ca7dd48c4e64145823055b229a350472\": container with ID starting with d95d50ae17d040d4b82dee640807bcf3ca7dd48c4e64145823055b229a350472 not found: ID does not exist" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.537037 4884 scope.go:117] "RemoveContainer" containerID="8639902764e9d3e836acbb0db3ade4689d085e3b79410839612d18b4f6d4545b" Dec 02 02:59:15 crc kubenswrapper[4884]: E1202 02:59:15.537373 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8639902764e9d3e836acbb0db3ade4689d085e3b79410839612d18b4f6d4545b\": container with ID starting with 8639902764e9d3e836acbb0db3ade4689d085e3b79410839612d18b4f6d4545b not found: ID does not exist" containerID="8639902764e9d3e836acbb0db3ade4689d085e3b79410839612d18b4f6d4545b" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.537393 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8639902764e9d3e836acbb0db3ade4689d085e3b79410839612d18b4f6d4545b"} err="failed to get container status \"8639902764e9d3e836acbb0db3ade4689d085e3b79410839612d18b4f6d4545b\": rpc error: code = NotFound desc = could not find container \"8639902764e9d3e836acbb0db3ade4689d085e3b79410839612d18b4f6d4545b\": container with ID starting with 8639902764e9d3e836acbb0db3ade4689d085e3b79410839612d18b4f6d4545b not found: ID does not exist" Dec 02 02:59:15 crc kubenswrapper[4884]: I1202 02:59:15.640952 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74c85300-3f81-4a20-8dd5-a8a95816e132" path="/var/lib/kubelet/pods/74c85300-3f81-4a20-8dd5-a8a95816e132/volumes" Dec 02 02:59:22 crc kubenswrapper[4884]: I1202 02:59:22.559334 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:22 crc kubenswrapper[4884]: I1202 02:59:22.612902 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:22 crc kubenswrapper[4884]: I1202 02:59:22.803605 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7bx8"] Dec 02 02:59:24 crc kubenswrapper[4884]: I1202 02:59:24.535488 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x7bx8" podUID="5ab5af5b-5b74-44af-a0bc-698aa76a47d4" containerName="registry-server" containerID="cri-o://82a3a2434948e7a4caa31f6fdc5ad3427ce0230f9a9cf4e8a947d5e0bc953ffa" gracePeriod=2 Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.154360 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.335364 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8l4th\" (UniqueName: \"kubernetes.io/projected/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-kube-api-access-8l4th\") pod \"5ab5af5b-5b74-44af-a0bc-698aa76a47d4\" (UID: \"5ab5af5b-5b74-44af-a0bc-698aa76a47d4\") " Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.335482 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-catalog-content\") pod \"5ab5af5b-5b74-44af-a0bc-698aa76a47d4\" (UID: \"5ab5af5b-5b74-44af-a0bc-698aa76a47d4\") " Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.335551 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-utilities\") pod \"5ab5af5b-5b74-44af-a0bc-698aa76a47d4\" (UID: \"5ab5af5b-5b74-44af-a0bc-698aa76a47d4\") " Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.336610 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-utilities" (OuterVolumeSpecName: "utilities") pod "5ab5af5b-5b74-44af-a0bc-698aa76a47d4" (UID: "5ab5af5b-5b74-44af-a0bc-698aa76a47d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.341884 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-kube-api-access-8l4th" (OuterVolumeSpecName: "kube-api-access-8l4th") pod "5ab5af5b-5b74-44af-a0bc-698aa76a47d4" (UID: "5ab5af5b-5b74-44af-a0bc-698aa76a47d4"). InnerVolumeSpecName "kube-api-access-8l4th". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.438717 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8l4th\" (UniqueName: \"kubernetes.io/projected/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-kube-api-access-8l4th\") on node \"crc\" DevicePath \"\"" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.438968 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.470855 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5ab5af5b-5b74-44af-a0bc-698aa76a47d4" (UID: "5ab5af5b-5b74-44af-a0bc-698aa76a47d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.541550 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5ab5af5b-5b74-44af-a0bc-698aa76a47d4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.565534 4884 generic.go:334] "Generic (PLEG): container finished" podID="5ab5af5b-5b74-44af-a0bc-698aa76a47d4" containerID="82a3a2434948e7a4caa31f6fdc5ad3427ce0230f9a9cf4e8a947d5e0bc953ffa" exitCode=0 Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.565640 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7bx8" event={"ID":"5ab5af5b-5b74-44af-a0bc-698aa76a47d4","Type":"ContainerDied","Data":"82a3a2434948e7a4caa31f6fdc5ad3427ce0230f9a9cf4e8a947d5e0bc953ffa"} Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.565889 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x7bx8" event={"ID":"5ab5af5b-5b74-44af-a0bc-698aa76a47d4","Type":"ContainerDied","Data":"216051586d6ff0fd905e787ea529e39e29692827c2b85f6cc423df276f6d85bf"} Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.565677 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x7bx8" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.565931 4884 scope.go:117] "RemoveContainer" containerID="82a3a2434948e7a4caa31f6fdc5ad3427ce0230f9a9cf4e8a947d5e0bc953ffa" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.620441 4884 scope.go:117] "RemoveContainer" containerID="6b39bd32afeabd6ccef761a109b5ffd91572d43ef1bf68ce77dee227ff72841c" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.637866 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x7bx8"] Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.648517 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x7bx8"] Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.664846 4884 scope.go:117] "RemoveContainer" containerID="0c7ceed92bb9427b4d57fbc5cc29d437e8bc2e50adeb81fd6817632d940f8158" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.735691 4884 scope.go:117] "RemoveContainer" containerID="82a3a2434948e7a4caa31f6fdc5ad3427ce0230f9a9cf4e8a947d5e0bc953ffa" Dec 02 02:59:25 crc kubenswrapper[4884]: E1202 02:59:25.736409 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82a3a2434948e7a4caa31f6fdc5ad3427ce0230f9a9cf4e8a947d5e0bc953ffa\": container with ID starting with 82a3a2434948e7a4caa31f6fdc5ad3427ce0230f9a9cf4e8a947d5e0bc953ffa not found: ID does not exist" containerID="82a3a2434948e7a4caa31f6fdc5ad3427ce0230f9a9cf4e8a947d5e0bc953ffa" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.736460 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82a3a2434948e7a4caa31f6fdc5ad3427ce0230f9a9cf4e8a947d5e0bc953ffa"} err="failed to get container status \"82a3a2434948e7a4caa31f6fdc5ad3427ce0230f9a9cf4e8a947d5e0bc953ffa\": rpc error: code = NotFound desc = could not find container \"82a3a2434948e7a4caa31f6fdc5ad3427ce0230f9a9cf4e8a947d5e0bc953ffa\": container with ID starting with 82a3a2434948e7a4caa31f6fdc5ad3427ce0230f9a9cf4e8a947d5e0bc953ffa not found: ID does not exist" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.736496 4884 scope.go:117] "RemoveContainer" containerID="6b39bd32afeabd6ccef761a109b5ffd91572d43ef1bf68ce77dee227ff72841c" Dec 02 02:59:25 crc kubenswrapper[4884]: E1202 02:59:25.737042 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b39bd32afeabd6ccef761a109b5ffd91572d43ef1bf68ce77dee227ff72841c\": container with ID starting with 6b39bd32afeabd6ccef761a109b5ffd91572d43ef1bf68ce77dee227ff72841c not found: ID does not exist" containerID="6b39bd32afeabd6ccef761a109b5ffd91572d43ef1bf68ce77dee227ff72841c" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.737087 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b39bd32afeabd6ccef761a109b5ffd91572d43ef1bf68ce77dee227ff72841c"} err="failed to get container status \"6b39bd32afeabd6ccef761a109b5ffd91572d43ef1bf68ce77dee227ff72841c\": rpc error: code = NotFound desc = could not find container \"6b39bd32afeabd6ccef761a109b5ffd91572d43ef1bf68ce77dee227ff72841c\": container with ID starting with 6b39bd32afeabd6ccef761a109b5ffd91572d43ef1bf68ce77dee227ff72841c not found: ID does not exist" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.737119 4884 scope.go:117] "RemoveContainer" containerID="0c7ceed92bb9427b4d57fbc5cc29d437e8bc2e50adeb81fd6817632d940f8158" Dec 02 02:59:25 crc kubenswrapper[4884]: E1202 02:59:25.737766 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c7ceed92bb9427b4d57fbc5cc29d437e8bc2e50adeb81fd6817632d940f8158\": container with ID starting with 0c7ceed92bb9427b4d57fbc5cc29d437e8bc2e50adeb81fd6817632d940f8158 not found: ID does not exist" containerID="0c7ceed92bb9427b4d57fbc5cc29d437e8bc2e50adeb81fd6817632d940f8158" Dec 02 02:59:25 crc kubenswrapper[4884]: I1202 02:59:25.737818 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c7ceed92bb9427b4d57fbc5cc29d437e8bc2e50adeb81fd6817632d940f8158"} err="failed to get container status \"0c7ceed92bb9427b4d57fbc5cc29d437e8bc2e50adeb81fd6817632d940f8158\": rpc error: code = NotFound desc = could not find container \"0c7ceed92bb9427b4d57fbc5cc29d437e8bc2e50adeb81fd6817632d940f8158\": container with ID starting with 0c7ceed92bb9427b4d57fbc5cc29d437e8bc2e50adeb81fd6817632d940f8158 not found: ID does not exist" Dec 02 02:59:27 crc kubenswrapper[4884]: I1202 02:59:27.631984 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ab5af5b-5b74-44af-a0bc-698aa76a47d4" path="/var/lib/kubelet/pods/5ab5af5b-5b74-44af-a0bc-698aa76a47d4/volumes" Dec 02 02:59:46 crc kubenswrapper[4884]: I1202 02:59:46.971067 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 02:59:46 crc kubenswrapper[4884]: I1202 02:59:46.971574 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.177639 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh"] Dec 02 03:00:00 crc kubenswrapper[4884]: E1202 03:00:00.179138 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c85300-3f81-4a20-8dd5-a8a95816e132" containerName="extract-utilities" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.179171 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c85300-3f81-4a20-8dd5-a8a95816e132" containerName="extract-utilities" Dec 02 03:00:00 crc kubenswrapper[4884]: E1202 03:00:00.179295 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c85300-3f81-4a20-8dd5-a8a95816e132" containerName="extract-content" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.179331 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c85300-3f81-4a20-8dd5-a8a95816e132" containerName="extract-content" Dec 02 03:00:00 crc kubenswrapper[4884]: E1202 03:00:00.179368 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab5af5b-5b74-44af-a0bc-698aa76a47d4" containerName="extract-content" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.179378 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab5af5b-5b74-44af-a0bc-698aa76a47d4" containerName="extract-content" Dec 02 03:00:00 crc kubenswrapper[4884]: E1202 03:00:00.179423 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab5af5b-5b74-44af-a0bc-698aa76a47d4" containerName="registry-server" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.179433 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab5af5b-5b74-44af-a0bc-698aa76a47d4" containerName="registry-server" Dec 02 03:00:00 crc kubenswrapper[4884]: E1202 03:00:00.179456 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab5af5b-5b74-44af-a0bc-698aa76a47d4" containerName="extract-utilities" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.179466 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab5af5b-5b74-44af-a0bc-698aa76a47d4" containerName="extract-utilities" Dec 02 03:00:00 crc kubenswrapper[4884]: E1202 03:00:00.179505 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74c85300-3f81-4a20-8dd5-a8a95816e132" containerName="registry-server" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.179513 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="74c85300-3f81-4a20-8dd5-a8a95816e132" containerName="registry-server" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.179960 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="74c85300-3f81-4a20-8dd5-a8a95816e132" containerName="registry-server" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.179988 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ab5af5b-5b74-44af-a0bc-698aa76a47d4" containerName="registry-server" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.180894 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.183575 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.184337 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.200288 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh"] Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.286345 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvhx4\" (UniqueName: \"kubernetes.io/projected/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-kube-api-access-wvhx4\") pod \"collect-profiles-29410740-m9dfh\" (UID: \"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.286445 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-secret-volume\") pod \"collect-profiles-29410740-m9dfh\" (UID: \"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.286624 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-config-volume\") pod \"collect-profiles-29410740-m9dfh\" (UID: \"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.389801 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvhx4\" (UniqueName: \"kubernetes.io/projected/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-kube-api-access-wvhx4\") pod \"collect-profiles-29410740-m9dfh\" (UID: \"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.389943 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-secret-volume\") pod \"collect-profiles-29410740-m9dfh\" (UID: \"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.390214 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-config-volume\") pod \"collect-profiles-29410740-m9dfh\" (UID: \"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.391588 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-config-volume\") pod \"collect-profiles-29410740-m9dfh\" (UID: \"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.400111 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-secret-volume\") pod \"collect-profiles-29410740-m9dfh\" (UID: \"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.414364 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvhx4\" (UniqueName: \"kubernetes.io/projected/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-kube-api-access-wvhx4\") pod \"collect-profiles-29410740-m9dfh\" (UID: \"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" Dec 02 03:00:00 crc kubenswrapper[4884]: I1202 03:00:00.527596 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" Dec 02 03:00:01 crc kubenswrapper[4884]: I1202 03:00:01.022343 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh"] Dec 02 03:00:02 crc kubenswrapper[4884]: I1202 03:00:02.064783 4884 generic.go:334] "Generic (PLEG): container finished" podID="cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4" containerID="416896c487aecb45a55109aeacec6829a43764365602e1bd82dd733c326a419e" exitCode=0 Dec 02 03:00:02 crc kubenswrapper[4884]: I1202 03:00:02.065090 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" event={"ID":"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4","Type":"ContainerDied","Data":"416896c487aecb45a55109aeacec6829a43764365602e1bd82dd733c326a419e"} Dec 02 03:00:02 crc kubenswrapper[4884]: I1202 03:00:02.065586 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" event={"ID":"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4","Type":"ContainerStarted","Data":"f1897bbcc27857091f4ee5ed7201de65ecbb432240cf4266969978fa4ff29f36"} Dec 02 03:00:03 crc kubenswrapper[4884]: I1202 03:00:03.494564 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" Dec 02 03:00:03 crc kubenswrapper[4884]: I1202 03:00:03.667054 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-secret-volume\") pod \"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4\" (UID: \"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4\") " Dec 02 03:00:03 crc kubenswrapper[4884]: I1202 03:00:03.667389 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-config-volume\") pod \"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4\" (UID: \"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4\") " Dec 02 03:00:03 crc kubenswrapper[4884]: I1202 03:00:03.667414 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvhx4\" (UniqueName: \"kubernetes.io/projected/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-kube-api-access-wvhx4\") pod \"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4\" (UID: \"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4\") " Dec 02 03:00:03 crc kubenswrapper[4884]: I1202 03:00:03.668247 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-config-volume" (OuterVolumeSpecName: "config-volume") pod "cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4" (UID: "cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 03:00:03 crc kubenswrapper[4884]: I1202 03:00:03.674449 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4" (UID: "cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 03:00:03 crc kubenswrapper[4884]: I1202 03:00:03.674973 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-kube-api-access-wvhx4" (OuterVolumeSpecName: "kube-api-access-wvhx4") pod "cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4" (UID: "cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4"). InnerVolumeSpecName "kube-api-access-wvhx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 03:00:03 crc kubenswrapper[4884]: I1202 03:00:03.770729 4884 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 03:00:03 crc kubenswrapper[4884]: I1202 03:00:03.771559 4884 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 03:00:03 crc kubenswrapper[4884]: I1202 03:00:03.771590 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvhx4\" (UniqueName: \"kubernetes.io/projected/cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4-kube-api-access-wvhx4\") on node \"crc\" DevicePath \"\"" Dec 02 03:00:04 crc kubenswrapper[4884]: I1202 03:00:04.092831 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" event={"ID":"cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4","Type":"ContainerDied","Data":"f1897bbcc27857091f4ee5ed7201de65ecbb432240cf4266969978fa4ff29f36"} Dec 02 03:00:04 crc kubenswrapper[4884]: I1202 03:00:04.093012 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1897bbcc27857091f4ee5ed7201de65ecbb432240cf4266969978fa4ff29f36" Dec 02 03:00:04 crc kubenswrapper[4884]: I1202 03:00:04.092934 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410740-m9dfh" Dec 02 03:00:04 crc kubenswrapper[4884]: I1202 03:00:04.607310 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v"] Dec 02 03:00:04 crc kubenswrapper[4884]: I1202 03:00:04.624575 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410695-t4d8v"] Dec 02 03:00:05 crc kubenswrapper[4884]: I1202 03:00:05.636362 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14a15ad2-77e7-4021-bab0-0594850ba5bb" path="/var/lib/kubelet/pods/14a15ad2-77e7-4021-bab0-0594850ba5bb/volumes" Dec 02 03:00:10 crc kubenswrapper[4884]: I1202 03:00:10.997358 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9cxb8"] Dec 02 03:00:10 crc kubenswrapper[4884]: E1202 03:00:10.998759 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4" containerName="collect-profiles" Dec 02 03:00:10 crc kubenswrapper[4884]: I1202 03:00:10.998781 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4" containerName="collect-profiles" Dec 02 03:00:10 crc kubenswrapper[4884]: I1202 03:00:10.999072 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbe0cf1b-4123-4ac0-b269-8fdbe0495bc4" containerName="collect-profiles" Dec 02 03:00:11 crc kubenswrapper[4884]: I1202 03:00:11.001301 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:11 crc kubenswrapper[4884]: I1202 03:00:11.027317 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9cxb8"] Dec 02 03:00:11 crc kubenswrapper[4884]: I1202 03:00:11.152223 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trd6w\" (UniqueName: \"kubernetes.io/projected/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-kube-api-access-trd6w\") pod \"certified-operators-9cxb8\" (UID: \"bb53dd71-b4c7-4af9-80c6-6d43cda5c888\") " pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:11 crc kubenswrapper[4884]: I1202 03:00:11.152380 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-catalog-content\") pod \"certified-operators-9cxb8\" (UID: \"bb53dd71-b4c7-4af9-80c6-6d43cda5c888\") " pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:11 crc kubenswrapper[4884]: I1202 03:00:11.152437 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-utilities\") pod \"certified-operators-9cxb8\" (UID: \"bb53dd71-b4c7-4af9-80c6-6d43cda5c888\") " pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:11 crc kubenswrapper[4884]: I1202 03:00:11.254382 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-catalog-content\") pod \"certified-operators-9cxb8\" (UID: \"bb53dd71-b4c7-4af9-80c6-6d43cda5c888\") " pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:11 crc kubenswrapper[4884]: I1202 03:00:11.254443 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-utilities\") pod \"certified-operators-9cxb8\" (UID: \"bb53dd71-b4c7-4af9-80c6-6d43cda5c888\") " pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:11 crc kubenswrapper[4884]: I1202 03:00:11.254530 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trd6w\" (UniqueName: \"kubernetes.io/projected/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-kube-api-access-trd6w\") pod \"certified-operators-9cxb8\" (UID: \"bb53dd71-b4c7-4af9-80c6-6d43cda5c888\") " pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:11 crc kubenswrapper[4884]: I1202 03:00:11.255088 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-utilities\") pod \"certified-operators-9cxb8\" (UID: \"bb53dd71-b4c7-4af9-80c6-6d43cda5c888\") " pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:11 crc kubenswrapper[4884]: I1202 03:00:11.255182 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-catalog-content\") pod \"certified-operators-9cxb8\" (UID: \"bb53dd71-b4c7-4af9-80c6-6d43cda5c888\") " pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:11 crc kubenswrapper[4884]: I1202 03:00:11.551555 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trd6w\" (UniqueName: \"kubernetes.io/projected/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-kube-api-access-trd6w\") pod \"certified-operators-9cxb8\" (UID: \"bb53dd71-b4c7-4af9-80c6-6d43cda5c888\") " pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:11 crc kubenswrapper[4884]: I1202 03:00:11.641896 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:12 crc kubenswrapper[4884]: W1202 03:00:12.138694 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb53dd71_b4c7_4af9_80c6_6d43cda5c888.slice/crio-639ea42e96d3fb611f3457b6b6cfeca2a6f7d07f4cd0aa6d96140d32600471b1 WatchSource:0}: Error finding container 639ea42e96d3fb611f3457b6b6cfeca2a6f7d07f4cd0aa6d96140d32600471b1: Status 404 returned error can't find the container with id 639ea42e96d3fb611f3457b6b6cfeca2a6f7d07f4cd0aa6d96140d32600471b1 Dec 02 03:00:12 crc kubenswrapper[4884]: I1202 03:00:12.150548 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9cxb8"] Dec 02 03:00:12 crc kubenswrapper[4884]: I1202 03:00:12.206058 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cxb8" event={"ID":"bb53dd71-b4c7-4af9-80c6-6d43cda5c888","Type":"ContainerStarted","Data":"639ea42e96d3fb611f3457b6b6cfeca2a6f7d07f4cd0aa6d96140d32600471b1"} Dec 02 03:00:13 crc kubenswrapper[4884]: I1202 03:00:13.220685 4884 generic.go:334] "Generic (PLEG): container finished" podID="bb53dd71-b4c7-4af9-80c6-6d43cda5c888" containerID="351659a3c31c048eff09a9f68d87426811c27eaa0b570b7973d61f8088dd2794" exitCode=0 Dec 02 03:00:13 crc kubenswrapper[4884]: I1202 03:00:13.220775 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cxb8" event={"ID":"bb53dd71-b4c7-4af9-80c6-6d43cda5c888","Type":"ContainerDied","Data":"351659a3c31c048eff09a9f68d87426811c27eaa0b570b7973d61f8088dd2794"} Dec 02 03:00:14 crc kubenswrapper[4884]: I1202 03:00:14.244375 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cxb8" event={"ID":"bb53dd71-b4c7-4af9-80c6-6d43cda5c888","Type":"ContainerStarted","Data":"81c64bcd2ea8a6269f734ae61514d9ff261d15ae53536be660e18256e7fc262b"} Dec 02 03:00:15 crc kubenswrapper[4884]: I1202 03:00:15.261180 4884 generic.go:334] "Generic (PLEG): container finished" podID="bb53dd71-b4c7-4af9-80c6-6d43cda5c888" containerID="81c64bcd2ea8a6269f734ae61514d9ff261d15ae53536be660e18256e7fc262b" exitCode=0 Dec 02 03:00:15 crc kubenswrapper[4884]: I1202 03:00:15.261244 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cxb8" event={"ID":"bb53dd71-b4c7-4af9-80c6-6d43cda5c888","Type":"ContainerDied","Data":"81c64bcd2ea8a6269f734ae61514d9ff261d15ae53536be660e18256e7fc262b"} Dec 02 03:00:16 crc kubenswrapper[4884]: I1202 03:00:16.276557 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cxb8" event={"ID":"bb53dd71-b4c7-4af9-80c6-6d43cda5c888","Type":"ContainerStarted","Data":"209c6f0c0d045ba0695d9c8dde8814dc88e8c0163c790354da1eb3882595c429"} Dec 02 03:00:16 crc kubenswrapper[4884]: I1202 03:00:16.311152 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9cxb8" podStartSLOduration=3.699815331 podStartE2EDuration="6.311135001s" podCreationTimestamp="2025-12-02 03:00:10 +0000 UTC" firstStartedPulling="2025-12-02 03:00:13.224699971 +0000 UTC m=+5029.900536885" lastFinishedPulling="2025-12-02 03:00:15.836019661 +0000 UTC m=+5032.511856555" observedRunningTime="2025-12-02 03:00:16.302190382 +0000 UTC m=+5032.978027276" watchObservedRunningTime="2025-12-02 03:00:16.311135001 +0000 UTC m=+5032.986971895" Dec 02 03:00:16 crc kubenswrapper[4884]: I1202 03:00:16.971294 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 03:00:16 crc kubenswrapper[4884]: I1202 03:00:16.971369 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 03:00:21 crc kubenswrapper[4884]: I1202 03:00:21.643537 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:21 crc kubenswrapper[4884]: I1202 03:00:21.643963 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:21 crc kubenswrapper[4884]: I1202 03:00:21.713922 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:22 crc kubenswrapper[4884]: I1202 03:00:22.437586 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:22 crc kubenswrapper[4884]: I1202 03:00:22.503297 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9cxb8"] Dec 02 03:00:24 crc kubenswrapper[4884]: I1202 03:00:24.376102 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9cxb8" podUID="bb53dd71-b4c7-4af9-80c6-6d43cda5c888" containerName="registry-server" containerID="cri-o://209c6f0c0d045ba0695d9c8dde8814dc88e8c0163c790354da1eb3882595c429" gracePeriod=2 Dec 02 03:00:24 crc kubenswrapper[4884]: I1202 03:00:24.975682 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.068362 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-utilities\") pod \"bb53dd71-b4c7-4af9-80c6-6d43cda5c888\" (UID: \"bb53dd71-b4c7-4af9-80c6-6d43cda5c888\") " Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.068799 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trd6w\" (UniqueName: \"kubernetes.io/projected/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-kube-api-access-trd6w\") pod \"bb53dd71-b4c7-4af9-80c6-6d43cda5c888\" (UID: \"bb53dd71-b4c7-4af9-80c6-6d43cda5c888\") " Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.068870 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-catalog-content\") pod \"bb53dd71-b4c7-4af9-80c6-6d43cda5c888\" (UID: \"bb53dd71-b4c7-4af9-80c6-6d43cda5c888\") " Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.069707 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-utilities" (OuterVolumeSpecName: "utilities") pod "bb53dd71-b4c7-4af9-80c6-6d43cda5c888" (UID: "bb53dd71-b4c7-4af9-80c6-6d43cda5c888"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.076341 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-kube-api-access-trd6w" (OuterVolumeSpecName: "kube-api-access-trd6w") pod "bb53dd71-b4c7-4af9-80c6-6d43cda5c888" (UID: "bb53dd71-b4c7-4af9-80c6-6d43cda5c888"). InnerVolumeSpecName "kube-api-access-trd6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.129374 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb53dd71-b4c7-4af9-80c6-6d43cda5c888" (UID: "bb53dd71-b4c7-4af9-80c6-6d43cda5c888"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.171103 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.171367 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trd6w\" (UniqueName: \"kubernetes.io/projected/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-kube-api-access-trd6w\") on node \"crc\" DevicePath \"\"" Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.171379 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb53dd71-b4c7-4af9-80c6-6d43cda5c888-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.391137 4884 generic.go:334] "Generic (PLEG): container finished" podID="bb53dd71-b4c7-4af9-80c6-6d43cda5c888" containerID="209c6f0c0d045ba0695d9c8dde8814dc88e8c0163c790354da1eb3882595c429" exitCode=0 Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.391454 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cxb8" event={"ID":"bb53dd71-b4c7-4af9-80c6-6d43cda5c888","Type":"ContainerDied","Data":"209c6f0c0d045ba0695d9c8dde8814dc88e8c0163c790354da1eb3882595c429"} Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.391502 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9cxb8" Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.391882 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9cxb8" event={"ID":"bb53dd71-b4c7-4af9-80c6-6d43cda5c888","Type":"ContainerDied","Data":"639ea42e96d3fb611f3457b6b6cfeca2a6f7d07f4cd0aa6d96140d32600471b1"} Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.391910 4884 scope.go:117] "RemoveContainer" containerID="209c6f0c0d045ba0695d9c8dde8814dc88e8c0163c790354da1eb3882595c429" Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.421962 4884 scope.go:117] "RemoveContainer" containerID="81c64bcd2ea8a6269f734ae61514d9ff261d15ae53536be660e18256e7fc262b" Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.464679 4884 scope.go:117] "RemoveContainer" containerID="351659a3c31c048eff09a9f68d87426811c27eaa0b570b7973d61f8088dd2794" Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.466118 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9cxb8"] Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.476541 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9cxb8"] Dec 02 03:00:25 crc kubenswrapper[4884]: I1202 03:00:25.628723 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb53dd71-b4c7-4af9-80c6-6d43cda5c888" path="/var/lib/kubelet/pods/bb53dd71-b4c7-4af9-80c6-6d43cda5c888/volumes" Dec 02 03:00:26 crc kubenswrapper[4884]: I1202 03:00:26.200589 4884 scope.go:117] "RemoveContainer" containerID="209c6f0c0d045ba0695d9c8dde8814dc88e8c0163c790354da1eb3882595c429" Dec 02 03:00:26 crc kubenswrapper[4884]: E1202 03:00:26.201127 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"209c6f0c0d045ba0695d9c8dde8814dc88e8c0163c790354da1eb3882595c429\": container with ID starting with 209c6f0c0d045ba0695d9c8dde8814dc88e8c0163c790354da1eb3882595c429 not found: ID does not exist" containerID="209c6f0c0d045ba0695d9c8dde8814dc88e8c0163c790354da1eb3882595c429" Dec 02 03:00:26 crc kubenswrapper[4884]: I1202 03:00:26.201180 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"209c6f0c0d045ba0695d9c8dde8814dc88e8c0163c790354da1eb3882595c429"} err="failed to get container status \"209c6f0c0d045ba0695d9c8dde8814dc88e8c0163c790354da1eb3882595c429\": rpc error: code = NotFound desc = could not find container \"209c6f0c0d045ba0695d9c8dde8814dc88e8c0163c790354da1eb3882595c429\": container with ID starting with 209c6f0c0d045ba0695d9c8dde8814dc88e8c0163c790354da1eb3882595c429 not found: ID does not exist" Dec 02 03:00:26 crc kubenswrapper[4884]: I1202 03:00:26.201214 4884 scope.go:117] "RemoveContainer" containerID="81c64bcd2ea8a6269f734ae61514d9ff261d15ae53536be660e18256e7fc262b" Dec 02 03:00:26 crc kubenswrapper[4884]: E1202 03:00:26.201671 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81c64bcd2ea8a6269f734ae61514d9ff261d15ae53536be660e18256e7fc262b\": container with ID starting with 81c64bcd2ea8a6269f734ae61514d9ff261d15ae53536be660e18256e7fc262b not found: ID does not exist" containerID="81c64bcd2ea8a6269f734ae61514d9ff261d15ae53536be660e18256e7fc262b" Dec 02 03:00:26 crc kubenswrapper[4884]: I1202 03:00:26.201698 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81c64bcd2ea8a6269f734ae61514d9ff261d15ae53536be660e18256e7fc262b"} err="failed to get container status \"81c64bcd2ea8a6269f734ae61514d9ff261d15ae53536be660e18256e7fc262b\": rpc error: code = NotFound desc = could not find container \"81c64bcd2ea8a6269f734ae61514d9ff261d15ae53536be660e18256e7fc262b\": container with ID starting with 81c64bcd2ea8a6269f734ae61514d9ff261d15ae53536be660e18256e7fc262b not found: ID does not exist" Dec 02 03:00:26 crc kubenswrapper[4884]: I1202 03:00:26.201733 4884 scope.go:117] "RemoveContainer" containerID="351659a3c31c048eff09a9f68d87426811c27eaa0b570b7973d61f8088dd2794" Dec 02 03:00:26 crc kubenswrapper[4884]: E1202 03:00:26.202214 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"351659a3c31c048eff09a9f68d87426811c27eaa0b570b7973d61f8088dd2794\": container with ID starting with 351659a3c31c048eff09a9f68d87426811c27eaa0b570b7973d61f8088dd2794 not found: ID does not exist" containerID="351659a3c31c048eff09a9f68d87426811c27eaa0b570b7973d61f8088dd2794" Dec 02 03:00:26 crc kubenswrapper[4884]: I1202 03:00:26.202302 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"351659a3c31c048eff09a9f68d87426811c27eaa0b570b7973d61f8088dd2794"} err="failed to get container status \"351659a3c31c048eff09a9f68d87426811c27eaa0b570b7973d61f8088dd2794\": rpc error: code = NotFound desc = could not find container \"351659a3c31c048eff09a9f68d87426811c27eaa0b570b7973d61f8088dd2794\": container with ID starting with 351659a3c31c048eff09a9f68d87426811c27eaa0b570b7973d61f8088dd2794 not found: ID does not exist" Dec 02 03:00:46 crc kubenswrapper[4884]: I1202 03:00:46.971421 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 03:00:46 crc kubenswrapper[4884]: I1202 03:00:46.972213 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 03:00:46 crc kubenswrapper[4884]: I1202 03:00:46.972335 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 03:00:46 crc kubenswrapper[4884]: I1202 03:00:46.973504 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1d849541abc4fdea729a5ec715545cbcd4dffd71b8ffc714c87d3321a01e2b9a"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 03:00:46 crc kubenswrapper[4884]: I1202 03:00:46.973607 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://1d849541abc4fdea729a5ec715545cbcd4dffd71b8ffc714c87d3321a01e2b9a" gracePeriod=600 Dec 02 03:00:47 crc kubenswrapper[4884]: I1202 03:00:47.661120 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="1d849541abc4fdea729a5ec715545cbcd4dffd71b8ffc714c87d3321a01e2b9a" exitCode=0 Dec 02 03:00:47 crc kubenswrapper[4884]: I1202 03:00:47.661194 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"1d849541abc4fdea729a5ec715545cbcd4dffd71b8ffc714c87d3321a01e2b9a"} Dec 02 03:00:47 crc kubenswrapper[4884]: I1202 03:00:47.661642 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a"} Dec 02 03:00:47 crc kubenswrapper[4884]: I1202 03:00:47.661665 4884 scope.go:117] "RemoveContainer" containerID="ba7b74175806b4150b622a2e69ffa21c9b6de5e60e61c850702cf82094a6c89b" Dec 02 03:00:47 crc kubenswrapper[4884]: I1202 03:00:47.679796 4884 scope.go:117] "RemoveContainer" containerID="ee9cef5f1904c01a727b9d8fd751eee706e65cce1996424423f4b20d41f39a1f" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.201718 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29410741-ngbgf"] Dec 02 03:01:00 crc kubenswrapper[4884]: E1202 03:01:00.203196 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb53dd71-b4c7-4af9-80c6-6d43cda5c888" containerName="extract-utilities" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.203235 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb53dd71-b4c7-4af9-80c6-6d43cda5c888" containerName="extract-utilities" Dec 02 03:01:00 crc kubenswrapper[4884]: E1202 03:01:00.203261 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb53dd71-b4c7-4af9-80c6-6d43cda5c888" containerName="registry-server" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.203271 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb53dd71-b4c7-4af9-80c6-6d43cda5c888" containerName="registry-server" Dec 02 03:01:00 crc kubenswrapper[4884]: E1202 03:01:00.203321 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb53dd71-b4c7-4af9-80c6-6d43cda5c888" containerName="extract-content" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.203329 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb53dd71-b4c7-4af9-80c6-6d43cda5c888" containerName="extract-content" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.203651 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb53dd71-b4c7-4af9-80c6-6d43cda5c888" containerName="registry-server" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.204688 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.215669 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29410741-ngbgf"] Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.302507 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-fernet-keys\") pod \"keystone-cron-29410741-ngbgf\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.302604 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-config-data\") pod \"keystone-cron-29410741-ngbgf\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.302700 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvnnh\" (UniqueName: \"kubernetes.io/projected/9123c6f1-236e-4598-8309-9f6f3784613c-kube-api-access-bvnnh\") pod \"keystone-cron-29410741-ngbgf\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.302890 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-combined-ca-bundle\") pod \"keystone-cron-29410741-ngbgf\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.404718 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-fernet-keys\") pod \"keystone-cron-29410741-ngbgf\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.404835 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-config-data\") pod \"keystone-cron-29410741-ngbgf\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.404920 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvnnh\" (UniqueName: \"kubernetes.io/projected/9123c6f1-236e-4598-8309-9f6f3784613c-kube-api-access-bvnnh\") pod \"keystone-cron-29410741-ngbgf\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.404955 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-combined-ca-bundle\") pod \"keystone-cron-29410741-ngbgf\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.412657 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-fernet-keys\") pod \"keystone-cron-29410741-ngbgf\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.415423 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-combined-ca-bundle\") pod \"keystone-cron-29410741-ngbgf\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.415441 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-config-data\") pod \"keystone-cron-29410741-ngbgf\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.428301 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvnnh\" (UniqueName: \"kubernetes.io/projected/9123c6f1-236e-4598-8309-9f6f3784613c-kube-api-access-bvnnh\") pod \"keystone-cron-29410741-ngbgf\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:01:00 crc kubenswrapper[4884]: I1202 03:01:00.536320 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:01:01 crc kubenswrapper[4884]: I1202 03:01:01.085599 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29410741-ngbgf"] Dec 02 03:01:01 crc kubenswrapper[4884]: I1202 03:01:01.829011 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410741-ngbgf" event={"ID":"9123c6f1-236e-4598-8309-9f6f3784613c","Type":"ContainerStarted","Data":"077555bdfeede9789a271f28aad3bc41535d49b01c72586a193aad3c0bf005a7"} Dec 02 03:01:01 crc kubenswrapper[4884]: I1202 03:01:01.829375 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410741-ngbgf" event={"ID":"9123c6f1-236e-4598-8309-9f6f3784613c","Type":"ContainerStarted","Data":"209ebeac388264bd823a0e811e78b956b044baace63911f450275a289ff55908"} Dec 02 03:01:01 crc kubenswrapper[4884]: I1202 03:01:01.862389 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29410741-ngbgf" podStartSLOduration=1.8623678350000001 podStartE2EDuration="1.862367835s" podCreationTimestamp="2025-12-02 03:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 03:01:01.849934551 +0000 UTC m=+5078.525771475" watchObservedRunningTime="2025-12-02 03:01:01.862367835 +0000 UTC m=+5078.538204729" Dec 02 03:01:03 crc kubenswrapper[4884]: I1202 03:01:03.857246 4884 generic.go:334] "Generic (PLEG): container finished" podID="9123c6f1-236e-4598-8309-9f6f3784613c" containerID="077555bdfeede9789a271f28aad3bc41535d49b01c72586a193aad3c0bf005a7" exitCode=0 Dec 02 03:01:03 crc kubenswrapper[4884]: I1202 03:01:03.857357 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410741-ngbgf" event={"ID":"9123c6f1-236e-4598-8309-9f6f3784613c","Type":"ContainerDied","Data":"077555bdfeede9789a271f28aad3bc41535d49b01c72586a193aad3c0bf005a7"} Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.250056 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.315784 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-fernet-keys\") pod \"9123c6f1-236e-4598-8309-9f6f3784613c\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.315970 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-combined-ca-bundle\") pod \"9123c6f1-236e-4598-8309-9f6f3784613c\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.316006 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-config-data\") pod \"9123c6f1-236e-4598-8309-9f6f3784613c\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.316119 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvnnh\" (UniqueName: \"kubernetes.io/projected/9123c6f1-236e-4598-8309-9f6f3784613c-kube-api-access-bvnnh\") pod \"9123c6f1-236e-4598-8309-9f6f3784613c\" (UID: \"9123c6f1-236e-4598-8309-9f6f3784613c\") " Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.324921 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9123c6f1-236e-4598-8309-9f6f3784613c" (UID: "9123c6f1-236e-4598-8309-9f6f3784613c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.329735 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9123c6f1-236e-4598-8309-9f6f3784613c-kube-api-access-bvnnh" (OuterVolumeSpecName: "kube-api-access-bvnnh") pod "9123c6f1-236e-4598-8309-9f6f3784613c" (UID: "9123c6f1-236e-4598-8309-9f6f3784613c"). InnerVolumeSpecName "kube-api-access-bvnnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.355650 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9123c6f1-236e-4598-8309-9f6f3784613c" (UID: "9123c6f1-236e-4598-8309-9f6f3784613c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.412452 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-config-data" (OuterVolumeSpecName: "config-data") pod "9123c6f1-236e-4598-8309-9f6f3784613c" (UID: "9123c6f1-236e-4598-8309-9f6f3784613c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.418485 4884 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.418511 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.418523 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvnnh\" (UniqueName: \"kubernetes.io/projected/9123c6f1-236e-4598-8309-9f6f3784613c-kube-api-access-bvnnh\") on node \"crc\" DevicePath \"\"" Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.418532 4884 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9123c6f1-236e-4598-8309-9f6f3784613c-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.885033 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29410741-ngbgf" event={"ID":"9123c6f1-236e-4598-8309-9f6f3784613c","Type":"ContainerDied","Data":"209ebeac388264bd823a0e811e78b956b044baace63911f450275a289ff55908"} Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.885077 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="209ebeac388264bd823a0e811e78b956b044baace63911f450275a289ff55908" Dec 02 03:01:05 crc kubenswrapper[4884]: I1202 03:01:05.885116 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29410741-ngbgf" Dec 02 03:03:16 crc kubenswrapper[4884]: I1202 03:03:16.972407 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 03:03:16 crc kubenswrapper[4884]: I1202 03:03:16.973409 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 03:03:46 crc kubenswrapper[4884]: I1202 03:03:46.971831 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 03:03:46 crc kubenswrapper[4884]: I1202 03:03:46.972723 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 03:04:16 crc kubenswrapper[4884]: I1202 03:04:16.971878 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 03:04:16 crc kubenswrapper[4884]: I1202 03:04:16.972620 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 03:04:16 crc kubenswrapper[4884]: I1202 03:04:16.972695 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 03:04:16 crc kubenswrapper[4884]: I1202 03:04:16.973817 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 03:04:16 crc kubenswrapper[4884]: I1202 03:04:16.973914 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" gracePeriod=600 Dec 02 03:04:17 crc kubenswrapper[4884]: E1202 03:04:17.103883 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:04:18 crc kubenswrapper[4884]: I1202 03:04:18.119049 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" exitCode=0 Dec 02 03:04:18 crc kubenswrapper[4884]: I1202 03:04:18.119093 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a"} Dec 02 03:04:18 crc kubenswrapper[4884]: I1202 03:04:18.119161 4884 scope.go:117] "RemoveContainer" containerID="1d849541abc4fdea729a5ec715545cbcd4dffd71b8ffc714c87d3321a01e2b9a" Dec 02 03:04:18 crc kubenswrapper[4884]: I1202 03:04:18.121341 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:04:18 crc kubenswrapper[4884]: E1202 03:04:18.121968 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:04:31 crc kubenswrapper[4884]: I1202 03:04:31.615909 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:04:31 crc kubenswrapper[4884]: E1202 03:04:31.617170 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:04:43 crc kubenswrapper[4884]: I1202 03:04:43.626219 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:04:43 crc kubenswrapper[4884]: E1202 03:04:43.627373 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:04:57 crc kubenswrapper[4884]: I1202 03:04:57.637837 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:04:57 crc kubenswrapper[4884]: E1202 03:04:57.638499 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:05:09 crc kubenswrapper[4884]: I1202 03:05:09.615424 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:05:09 crc kubenswrapper[4884]: E1202 03:05:09.616409 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:05:21 crc kubenswrapper[4884]: I1202 03:05:21.626707 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:05:21 crc kubenswrapper[4884]: E1202 03:05:21.627959 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:05:34 crc kubenswrapper[4884]: I1202 03:05:34.614529 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:05:34 crc kubenswrapper[4884]: E1202 03:05:34.616015 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:05:45 crc kubenswrapper[4884]: I1202 03:05:45.615091 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:05:45 crc kubenswrapper[4884]: E1202 03:05:45.616097 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:05:59 crc kubenswrapper[4884]: I1202 03:05:59.614650 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:05:59 crc kubenswrapper[4884]: E1202 03:05:59.615725 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:06:11 crc kubenswrapper[4884]: I1202 03:06:11.614594 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:06:11 crc kubenswrapper[4884]: E1202 03:06:11.615935 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:06:21 crc kubenswrapper[4884]: I1202 03:06:21.687834 4884 generic.go:334] "Generic (PLEG): container finished" podID="56afebdd-31ad-4882-b405-53ada83cbaea" containerID="9ed0eab2809a6d980eb0bd4cd922380106ecffb5c9cbaffa831c89315a5b8a9d" exitCode=1 Dec 02 03:06:21 crc kubenswrapper[4884]: I1202 03:06:21.688284 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"56afebdd-31ad-4882-b405-53ada83cbaea","Type":"ContainerDied","Data":"9ed0eab2809a6d980eb0bd4cd922380106ecffb5c9cbaffa831c89315a5b8a9d"} Dec 02 03:06:22 crc kubenswrapper[4884]: I1202 03:06:22.614343 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:06:22 crc kubenswrapper[4884]: E1202 03:06:22.615201 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.079205 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.171815 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-ca-certs\") pod \"56afebdd-31ad-4882-b405-53ada83cbaea\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.172025 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpbvf\" (UniqueName: \"kubernetes.io/projected/56afebdd-31ad-4882-b405-53ada83cbaea-kube-api-access-bpbvf\") pod \"56afebdd-31ad-4882-b405-53ada83cbaea\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.172076 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56afebdd-31ad-4882-b405-53ada83cbaea-config-data\") pod \"56afebdd-31ad-4882-b405-53ada83cbaea\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.172114 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-ssh-key\") pod \"56afebdd-31ad-4882-b405-53ada83cbaea\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.172171 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/56afebdd-31ad-4882-b405-53ada83cbaea-test-operator-ephemeral-temporary\") pod \"56afebdd-31ad-4882-b405-53ada83cbaea\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.172213 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"56afebdd-31ad-4882-b405-53ada83cbaea\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.172257 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-openstack-config-secret\") pod \"56afebdd-31ad-4882-b405-53ada83cbaea\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.172766 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/56afebdd-31ad-4882-b405-53ada83cbaea-openstack-config\") pod \"56afebdd-31ad-4882-b405-53ada83cbaea\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.172789 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/56afebdd-31ad-4882-b405-53ada83cbaea-test-operator-ephemeral-workdir\") pod \"56afebdd-31ad-4882-b405-53ada83cbaea\" (UID: \"56afebdd-31ad-4882-b405-53ada83cbaea\") " Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.172817 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56afebdd-31ad-4882-b405-53ada83cbaea-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "56afebdd-31ad-4882-b405-53ada83cbaea" (UID: "56afebdd-31ad-4882-b405-53ada83cbaea"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.173271 4884 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/56afebdd-31ad-4882-b405-53ada83cbaea-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.173450 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56afebdd-31ad-4882-b405-53ada83cbaea-config-data" (OuterVolumeSpecName: "config-data") pod "56afebdd-31ad-4882-b405-53ada83cbaea" (UID: "56afebdd-31ad-4882-b405-53ada83cbaea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.177509 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "56afebdd-31ad-4882-b405-53ada83cbaea" (UID: "56afebdd-31ad-4882-b405-53ada83cbaea"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.184899 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56afebdd-31ad-4882-b405-53ada83cbaea-kube-api-access-bpbvf" (OuterVolumeSpecName: "kube-api-access-bpbvf") pod "56afebdd-31ad-4882-b405-53ada83cbaea" (UID: "56afebdd-31ad-4882-b405-53ada83cbaea"). InnerVolumeSpecName "kube-api-access-bpbvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.207886 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "56afebdd-31ad-4882-b405-53ada83cbaea" (UID: "56afebdd-31ad-4882-b405-53ada83cbaea"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.217223 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "56afebdd-31ad-4882-b405-53ada83cbaea" (UID: "56afebdd-31ad-4882-b405-53ada83cbaea"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.222610 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "56afebdd-31ad-4882-b405-53ada83cbaea" (UID: "56afebdd-31ad-4882-b405-53ada83cbaea"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.231309 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56afebdd-31ad-4882-b405-53ada83cbaea-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "56afebdd-31ad-4882-b405-53ada83cbaea" (UID: "56afebdd-31ad-4882-b405-53ada83cbaea"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.251382 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56afebdd-31ad-4882-b405-53ada83cbaea-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "56afebdd-31ad-4882-b405-53ada83cbaea" (UID: "56afebdd-31ad-4882-b405-53ada83cbaea"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.274979 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpbvf\" (UniqueName: \"kubernetes.io/projected/56afebdd-31ad-4882-b405-53ada83cbaea-kube-api-access-bpbvf\") on node \"crc\" DevicePath \"\"" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.275016 4884 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/56afebdd-31ad-4882-b405-53ada83cbaea-config-data\") on node \"crc\" DevicePath \"\"" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.275029 4884 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.275068 4884 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.275081 4884 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.275092 4884 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/56afebdd-31ad-4882-b405-53ada83cbaea-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.275106 4884 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/56afebdd-31ad-4882-b405-53ada83cbaea-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.275119 4884 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/56afebdd-31ad-4882-b405-53ada83cbaea-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.294946 4884 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.377899 4884 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.708304 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"56afebdd-31ad-4882-b405-53ada83cbaea","Type":"ContainerDied","Data":"ea1f47aea8b46bc72b37b2869347d735dbe1e7654cc60ebd7edaf36f648094ba"} Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.708683 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea1f47aea8b46bc72b37b2869347d735dbe1e7654cc60ebd7edaf36f648094ba" Dec 02 03:06:23 crc kubenswrapper[4884]: I1202 03:06:23.708374 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 02 03:06:32 crc kubenswrapper[4884]: I1202 03:06:32.899570 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 02 03:06:32 crc kubenswrapper[4884]: E1202 03:06:32.900928 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9123c6f1-236e-4598-8309-9f6f3784613c" containerName="keystone-cron" Dec 02 03:06:32 crc kubenswrapper[4884]: I1202 03:06:32.900952 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="9123c6f1-236e-4598-8309-9f6f3784613c" containerName="keystone-cron" Dec 02 03:06:32 crc kubenswrapper[4884]: E1202 03:06:32.901018 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56afebdd-31ad-4882-b405-53ada83cbaea" containerName="tempest-tests-tempest-tests-runner" Dec 02 03:06:32 crc kubenswrapper[4884]: I1202 03:06:32.901043 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="56afebdd-31ad-4882-b405-53ada83cbaea" containerName="tempest-tests-tempest-tests-runner" Dec 02 03:06:32 crc kubenswrapper[4884]: I1202 03:06:32.901455 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="56afebdd-31ad-4882-b405-53ada83cbaea" containerName="tempest-tests-tempest-tests-runner" Dec 02 03:06:32 crc kubenswrapper[4884]: I1202 03:06:32.901515 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="9123c6f1-236e-4598-8309-9f6f3784613c" containerName="keystone-cron" Dec 02 03:06:32 crc kubenswrapper[4884]: I1202 03:06:32.903045 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 03:06:32 crc kubenswrapper[4884]: I1202 03:06:32.905216 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-v2jd6" Dec 02 03:06:32 crc kubenswrapper[4884]: I1202 03:06:32.913216 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 02 03:06:33 crc kubenswrapper[4884]: I1202 03:06:33.031945 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbtg9\" (UniqueName: \"kubernetes.io/projected/6cd1aa1b-940f-40ab-9776-f74981c420df-kube-api-access-nbtg9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"6cd1aa1b-940f-40ab-9776-f74981c420df\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 03:06:33 crc kubenswrapper[4884]: I1202 03:06:33.032182 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"6cd1aa1b-940f-40ab-9776-f74981c420df\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 03:06:33 crc kubenswrapper[4884]: I1202 03:06:33.133930 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"6cd1aa1b-940f-40ab-9776-f74981c420df\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 03:06:33 crc kubenswrapper[4884]: I1202 03:06:33.134340 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbtg9\" (UniqueName: \"kubernetes.io/projected/6cd1aa1b-940f-40ab-9776-f74981c420df-kube-api-access-nbtg9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"6cd1aa1b-940f-40ab-9776-f74981c420df\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 03:06:33 crc kubenswrapper[4884]: I1202 03:06:33.134484 4884 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"6cd1aa1b-940f-40ab-9776-f74981c420df\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 03:06:33 crc kubenswrapper[4884]: I1202 03:06:33.165118 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbtg9\" (UniqueName: \"kubernetes.io/projected/6cd1aa1b-940f-40ab-9776-f74981c420df-kube-api-access-nbtg9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"6cd1aa1b-940f-40ab-9776-f74981c420df\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 03:06:33 crc kubenswrapper[4884]: I1202 03:06:33.174921 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"6cd1aa1b-940f-40ab-9776-f74981c420df\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 03:06:33 crc kubenswrapper[4884]: I1202 03:06:33.228764 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 02 03:06:33 crc kubenswrapper[4884]: I1202 03:06:33.803203 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 02 03:06:33 crc kubenswrapper[4884]: W1202 03:06:33.815356 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6cd1aa1b_940f_40ab_9776_f74981c420df.slice/crio-be8854d15484d7ccd7c6120afc7fe5f4427c2221cb005b75a7f43d87c4262a7d WatchSource:0}: Error finding container be8854d15484d7ccd7c6120afc7fe5f4427c2221cb005b75a7f43d87c4262a7d: Status 404 returned error can't find the container with id be8854d15484d7ccd7c6120afc7fe5f4427c2221cb005b75a7f43d87c4262a7d Dec 02 03:06:33 crc kubenswrapper[4884]: I1202 03:06:33.821448 4884 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 03:06:34 crc kubenswrapper[4884]: I1202 03:06:34.841548 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"6cd1aa1b-940f-40ab-9776-f74981c420df","Type":"ContainerStarted","Data":"be8854d15484d7ccd7c6120afc7fe5f4427c2221cb005b75a7f43d87c4262a7d"} Dec 02 03:06:35 crc kubenswrapper[4884]: I1202 03:06:35.619357 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:06:35 crc kubenswrapper[4884]: E1202 03:06:35.620202 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:06:35 crc kubenswrapper[4884]: I1202 03:06:35.851145 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"6cd1aa1b-940f-40ab-9776-f74981c420df","Type":"ContainerStarted","Data":"ec78ceeeb78743188947618658dbf17561465c0db98538f12295cbce02feb28f"} Dec 02 03:06:35 crc kubenswrapper[4884]: I1202 03:06:35.869938 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=3.0313821 podStartE2EDuration="3.869923278s" podCreationTimestamp="2025-12-02 03:06:32 +0000 UTC" firstStartedPulling="2025-12-02 03:06:33.821074086 +0000 UTC m=+5410.496911000" lastFinishedPulling="2025-12-02 03:06:34.659615284 +0000 UTC m=+5411.335452178" observedRunningTime="2025-12-02 03:06:35.866148366 +0000 UTC m=+5412.541985250" watchObservedRunningTime="2025-12-02 03:06:35.869923278 +0000 UTC m=+5412.545760162" Dec 02 03:06:50 crc kubenswrapper[4884]: I1202 03:06:50.616424 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:06:50 crc kubenswrapper[4884]: E1202 03:06:50.617490 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:07:03 crc kubenswrapper[4884]: I1202 03:07:03.627914 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:07:03 crc kubenswrapper[4884]: E1202 03:07:03.629359 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:07:14 crc kubenswrapper[4884]: I1202 03:07:14.614916 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:07:14 crc kubenswrapper[4884]: E1202 03:07:14.615975 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:07:14 crc kubenswrapper[4884]: I1202 03:07:14.937834 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h4s98/must-gather-vwhm4"] Dec 02 03:07:14 crc kubenswrapper[4884]: I1202 03:07:14.940490 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/must-gather-vwhm4" Dec 02 03:07:14 crc kubenswrapper[4884]: I1202 03:07:14.942381 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-h4s98"/"default-dockercfg-27vx2" Dec 02 03:07:14 crc kubenswrapper[4884]: I1202 03:07:14.942978 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-h4s98"/"kube-root-ca.crt" Dec 02 03:07:14 crc kubenswrapper[4884]: I1202 03:07:14.943013 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-h4s98"/"openshift-service-ca.crt" Dec 02 03:07:14 crc kubenswrapper[4884]: I1202 03:07:14.959358 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h4s98/must-gather-vwhm4"] Dec 02 03:07:15 crc kubenswrapper[4884]: I1202 03:07:15.028158 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/284a08f7-5dcd-48f1-823e-2c4b4e8934da-must-gather-output\") pod \"must-gather-vwhm4\" (UID: \"284a08f7-5dcd-48f1-823e-2c4b4e8934da\") " pod="openshift-must-gather-h4s98/must-gather-vwhm4" Dec 02 03:07:15 crc kubenswrapper[4884]: I1202 03:07:15.028457 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zdct\" (UniqueName: \"kubernetes.io/projected/284a08f7-5dcd-48f1-823e-2c4b4e8934da-kube-api-access-4zdct\") pod \"must-gather-vwhm4\" (UID: \"284a08f7-5dcd-48f1-823e-2c4b4e8934da\") " pod="openshift-must-gather-h4s98/must-gather-vwhm4" Dec 02 03:07:15 crc kubenswrapper[4884]: I1202 03:07:15.130035 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/284a08f7-5dcd-48f1-823e-2c4b4e8934da-must-gather-output\") pod \"must-gather-vwhm4\" (UID: \"284a08f7-5dcd-48f1-823e-2c4b4e8934da\") " pod="openshift-must-gather-h4s98/must-gather-vwhm4" Dec 02 03:07:15 crc kubenswrapper[4884]: I1202 03:07:15.130605 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/284a08f7-5dcd-48f1-823e-2c4b4e8934da-must-gather-output\") pod \"must-gather-vwhm4\" (UID: \"284a08f7-5dcd-48f1-823e-2c4b4e8934da\") " pod="openshift-must-gather-h4s98/must-gather-vwhm4" Dec 02 03:07:15 crc kubenswrapper[4884]: I1202 03:07:15.130811 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zdct\" (UniqueName: \"kubernetes.io/projected/284a08f7-5dcd-48f1-823e-2c4b4e8934da-kube-api-access-4zdct\") pod \"must-gather-vwhm4\" (UID: \"284a08f7-5dcd-48f1-823e-2c4b4e8934da\") " pod="openshift-must-gather-h4s98/must-gather-vwhm4" Dec 02 03:07:15 crc kubenswrapper[4884]: I1202 03:07:15.150438 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zdct\" (UniqueName: \"kubernetes.io/projected/284a08f7-5dcd-48f1-823e-2c4b4e8934da-kube-api-access-4zdct\") pod \"must-gather-vwhm4\" (UID: \"284a08f7-5dcd-48f1-823e-2c4b4e8934da\") " pod="openshift-must-gather-h4s98/must-gather-vwhm4" Dec 02 03:07:15 crc kubenswrapper[4884]: I1202 03:07:15.261391 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/must-gather-vwhm4" Dec 02 03:07:15 crc kubenswrapper[4884]: I1202 03:07:15.760764 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h4s98/must-gather-vwhm4"] Dec 02 03:07:16 crc kubenswrapper[4884]: I1202 03:07:16.381548 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h4s98/must-gather-vwhm4" event={"ID":"284a08f7-5dcd-48f1-823e-2c4b4e8934da","Type":"ContainerStarted","Data":"408aea2322c5d4d44c16bd060d7839eb6d1bd41bf770e26d0516d59fdf24e86b"} Dec 02 03:07:21 crc kubenswrapper[4884]: I1202 03:07:21.441975 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h4s98/must-gather-vwhm4" event={"ID":"284a08f7-5dcd-48f1-823e-2c4b4e8934da","Type":"ContainerStarted","Data":"17508bad1f79c4bc670ee94895e9a60d9964781ab58bdd7355aa072eb25d5abd"} Dec 02 03:07:21 crc kubenswrapper[4884]: I1202 03:07:21.442608 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h4s98/must-gather-vwhm4" event={"ID":"284a08f7-5dcd-48f1-823e-2c4b4e8934da","Type":"ContainerStarted","Data":"1c40f6b01024689662e8040fd9156f4deff736d760bea7af7162c3b432ba8266"} Dec 02 03:07:21 crc kubenswrapper[4884]: I1202 03:07:21.467306 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-h4s98/must-gather-vwhm4" podStartSLOduration=2.62699492 podStartE2EDuration="7.467284125s" podCreationTimestamp="2025-12-02 03:07:14 +0000 UTC" firstStartedPulling="2025-12-02 03:07:15.793878245 +0000 UTC m=+5452.469715179" lastFinishedPulling="2025-12-02 03:07:20.63416747 +0000 UTC m=+5457.310004384" observedRunningTime="2025-12-02 03:07:21.46218026 +0000 UTC m=+5458.138017174" watchObservedRunningTime="2025-12-02 03:07:21.467284125 +0000 UTC m=+5458.143121029" Dec 02 03:07:24 crc kubenswrapper[4884]: I1202 03:07:24.569662 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h4s98/crc-debug-gtfgg"] Dec 02 03:07:24 crc kubenswrapper[4884]: I1202 03:07:24.571340 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/crc-debug-gtfgg" Dec 02 03:07:24 crc kubenswrapper[4884]: I1202 03:07:24.626595 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4-host\") pod \"crc-debug-gtfgg\" (UID: \"c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4\") " pod="openshift-must-gather-h4s98/crc-debug-gtfgg" Dec 02 03:07:24 crc kubenswrapper[4884]: I1202 03:07:24.626760 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmfb9\" (UniqueName: \"kubernetes.io/projected/c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4-kube-api-access-cmfb9\") pod \"crc-debug-gtfgg\" (UID: \"c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4\") " pod="openshift-must-gather-h4s98/crc-debug-gtfgg" Dec 02 03:07:24 crc kubenswrapper[4884]: I1202 03:07:24.728885 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4-host\") pod \"crc-debug-gtfgg\" (UID: \"c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4\") " pod="openshift-must-gather-h4s98/crc-debug-gtfgg" Dec 02 03:07:24 crc kubenswrapper[4884]: I1202 03:07:24.729076 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmfb9\" (UniqueName: \"kubernetes.io/projected/c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4-kube-api-access-cmfb9\") pod \"crc-debug-gtfgg\" (UID: \"c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4\") " pod="openshift-must-gather-h4s98/crc-debug-gtfgg" Dec 02 03:07:24 crc kubenswrapper[4884]: I1202 03:07:24.730740 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4-host\") pod \"crc-debug-gtfgg\" (UID: \"c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4\") " pod="openshift-must-gather-h4s98/crc-debug-gtfgg" Dec 02 03:07:24 crc kubenswrapper[4884]: I1202 03:07:24.755072 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmfb9\" (UniqueName: \"kubernetes.io/projected/c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4-kube-api-access-cmfb9\") pod \"crc-debug-gtfgg\" (UID: \"c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4\") " pod="openshift-must-gather-h4s98/crc-debug-gtfgg" Dec 02 03:07:24 crc kubenswrapper[4884]: I1202 03:07:24.887454 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/crc-debug-gtfgg" Dec 02 03:07:24 crc kubenswrapper[4884]: W1202 03:07:24.935382 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6f7b03b_3c92_44f7_8b4c_8c40f8a969f4.slice/crio-d7ebd44b59f6be846f96df1b873790e93791e6b717a1796cb1bf6b8962bd7b3f WatchSource:0}: Error finding container d7ebd44b59f6be846f96df1b873790e93791e6b717a1796cb1bf6b8962bd7b3f: Status 404 returned error can't find the container with id d7ebd44b59f6be846f96df1b873790e93791e6b717a1796cb1bf6b8962bd7b3f Dec 02 03:07:25 crc kubenswrapper[4884]: I1202 03:07:25.484154 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h4s98/crc-debug-gtfgg" event={"ID":"c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4","Type":"ContainerStarted","Data":"d7ebd44b59f6be846f96df1b873790e93791e6b717a1796cb1bf6b8962bd7b3f"} Dec 02 03:07:26 crc kubenswrapper[4884]: I1202 03:07:26.614961 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:07:26 crc kubenswrapper[4884]: E1202 03:07:26.615536 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:07:35 crc kubenswrapper[4884]: I1202 03:07:35.589258 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h4s98/crc-debug-gtfgg" event={"ID":"c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4","Type":"ContainerStarted","Data":"1fab3ed3787bf5044ebb631ddad31dbd559aac510caa4fd1512b6d81ea9d0abe"} Dec 02 03:07:35 crc kubenswrapper[4884]: I1202 03:07:35.606151 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-h4s98/crc-debug-gtfgg" podStartSLOduration=1.189064481 podStartE2EDuration="11.606137041s" podCreationTimestamp="2025-12-02 03:07:24 +0000 UTC" firstStartedPulling="2025-12-02 03:07:24.937825774 +0000 UTC m=+5461.613662658" lastFinishedPulling="2025-12-02 03:07:35.354898334 +0000 UTC m=+5472.030735218" observedRunningTime="2025-12-02 03:07:35.601784195 +0000 UTC m=+5472.277621079" watchObservedRunningTime="2025-12-02 03:07:35.606137041 +0000 UTC m=+5472.281973925" Dec 02 03:07:39 crc kubenswrapper[4884]: I1202 03:07:39.614713 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:07:39 crc kubenswrapper[4884]: E1202 03:07:39.615561 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:07:54 crc kubenswrapper[4884]: I1202 03:07:54.614422 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:07:54 crc kubenswrapper[4884]: E1202 03:07:54.615065 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:08:08 crc kubenswrapper[4884]: I1202 03:08:08.614975 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:08:08 crc kubenswrapper[4884]: E1202 03:08:08.616024 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:08:21 crc kubenswrapper[4884]: I1202 03:08:21.069909 4884 generic.go:334] "Generic (PLEG): container finished" podID="c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4" containerID="1fab3ed3787bf5044ebb631ddad31dbd559aac510caa4fd1512b6d81ea9d0abe" exitCode=0 Dec 02 03:08:21 crc kubenswrapper[4884]: I1202 03:08:21.070602 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h4s98/crc-debug-gtfgg" event={"ID":"c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4","Type":"ContainerDied","Data":"1fab3ed3787bf5044ebb631ddad31dbd559aac510caa4fd1512b6d81ea9d0abe"} Dec 02 03:08:21 crc kubenswrapper[4884]: I1202 03:08:21.616293 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:08:21 crc kubenswrapper[4884]: E1202 03:08:21.616782 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:08:22 crc kubenswrapper[4884]: I1202 03:08:22.183146 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/crc-debug-gtfgg" Dec 02 03:08:22 crc kubenswrapper[4884]: I1202 03:08:22.219877 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-h4s98/crc-debug-gtfgg"] Dec 02 03:08:22 crc kubenswrapper[4884]: I1202 03:08:22.230979 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-h4s98/crc-debug-gtfgg"] Dec 02 03:08:22 crc kubenswrapper[4884]: I1202 03:08:22.327804 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4-host\") pod \"c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4\" (UID: \"c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4\") " Dec 02 03:08:22 crc kubenswrapper[4884]: I1202 03:08:22.327943 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4-host" (OuterVolumeSpecName: "host") pod "c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4" (UID: "c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 03:08:22 crc kubenswrapper[4884]: I1202 03:08:22.328027 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmfb9\" (UniqueName: \"kubernetes.io/projected/c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4-kube-api-access-cmfb9\") pod \"c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4\" (UID: \"c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4\") " Dec 02 03:08:22 crc kubenswrapper[4884]: I1202 03:08:22.328503 4884 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4-host\") on node \"crc\" DevicePath \"\"" Dec 02 03:08:22 crc kubenswrapper[4884]: I1202 03:08:22.333944 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4-kube-api-access-cmfb9" (OuterVolumeSpecName: "kube-api-access-cmfb9") pod "c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4" (UID: "c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4"). InnerVolumeSpecName "kube-api-access-cmfb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 03:08:22 crc kubenswrapper[4884]: I1202 03:08:22.430763 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmfb9\" (UniqueName: \"kubernetes.io/projected/c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4-kube-api-access-cmfb9\") on node \"crc\" DevicePath \"\"" Dec 02 03:08:23 crc kubenswrapper[4884]: I1202 03:08:23.090567 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7ebd44b59f6be846f96df1b873790e93791e6b717a1796cb1bf6b8962bd7b3f" Dec 02 03:08:23 crc kubenswrapper[4884]: I1202 03:08:23.090661 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/crc-debug-gtfgg" Dec 02 03:08:23 crc kubenswrapper[4884]: I1202 03:08:23.490976 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h4s98/crc-debug-c9hgj"] Dec 02 03:08:23 crc kubenswrapper[4884]: E1202 03:08:23.493049 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4" containerName="container-00" Dec 02 03:08:23 crc kubenswrapper[4884]: I1202 03:08:23.493184 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4" containerName="container-00" Dec 02 03:08:23 crc kubenswrapper[4884]: I1202 03:08:23.493662 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4" containerName="container-00" Dec 02 03:08:23 crc kubenswrapper[4884]: I1202 03:08:23.494896 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/crc-debug-c9hgj" Dec 02 03:08:23 crc kubenswrapper[4884]: I1202 03:08:23.553108 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcpcb\" (UniqueName: \"kubernetes.io/projected/973d6331-20f4-4cce-a5c2-163e11f76d22-kube-api-access-bcpcb\") pod \"crc-debug-c9hgj\" (UID: \"973d6331-20f4-4cce-a5c2-163e11f76d22\") " pod="openshift-must-gather-h4s98/crc-debug-c9hgj" Dec 02 03:08:23 crc kubenswrapper[4884]: I1202 03:08:23.553169 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/973d6331-20f4-4cce-a5c2-163e11f76d22-host\") pod \"crc-debug-c9hgj\" (UID: \"973d6331-20f4-4cce-a5c2-163e11f76d22\") " pod="openshift-must-gather-h4s98/crc-debug-c9hgj" Dec 02 03:08:23 crc kubenswrapper[4884]: I1202 03:08:23.627465 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4" path="/var/lib/kubelet/pods/c6f7b03b-3c92-44f7-8b4c-8c40f8a969f4/volumes" Dec 02 03:08:23 crc kubenswrapper[4884]: I1202 03:08:23.655281 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcpcb\" (UniqueName: \"kubernetes.io/projected/973d6331-20f4-4cce-a5c2-163e11f76d22-kube-api-access-bcpcb\") pod \"crc-debug-c9hgj\" (UID: \"973d6331-20f4-4cce-a5c2-163e11f76d22\") " pod="openshift-must-gather-h4s98/crc-debug-c9hgj" Dec 02 03:08:23 crc kubenswrapper[4884]: I1202 03:08:23.655375 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/973d6331-20f4-4cce-a5c2-163e11f76d22-host\") pod \"crc-debug-c9hgj\" (UID: \"973d6331-20f4-4cce-a5c2-163e11f76d22\") " pod="openshift-must-gather-h4s98/crc-debug-c9hgj" Dec 02 03:08:23 crc kubenswrapper[4884]: I1202 03:08:23.655618 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/973d6331-20f4-4cce-a5c2-163e11f76d22-host\") pod \"crc-debug-c9hgj\" (UID: \"973d6331-20f4-4cce-a5c2-163e11f76d22\") " pod="openshift-must-gather-h4s98/crc-debug-c9hgj" Dec 02 03:08:23 crc kubenswrapper[4884]: I1202 03:08:23.675250 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcpcb\" (UniqueName: \"kubernetes.io/projected/973d6331-20f4-4cce-a5c2-163e11f76d22-kube-api-access-bcpcb\") pod \"crc-debug-c9hgj\" (UID: \"973d6331-20f4-4cce-a5c2-163e11f76d22\") " pod="openshift-must-gather-h4s98/crc-debug-c9hgj" Dec 02 03:08:23 crc kubenswrapper[4884]: I1202 03:08:23.816669 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/crc-debug-c9hgj" Dec 02 03:08:24 crc kubenswrapper[4884]: I1202 03:08:24.099335 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h4s98/crc-debug-c9hgj" event={"ID":"973d6331-20f4-4cce-a5c2-163e11f76d22","Type":"ContainerStarted","Data":"1032ec5163c2492db9cacf75bd0ea2eec65d5903dfa4b08fdb4d2ceb746c466f"} Dec 02 03:08:24 crc kubenswrapper[4884]: I1202 03:08:24.099625 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h4s98/crc-debug-c9hgj" event={"ID":"973d6331-20f4-4cce-a5c2-163e11f76d22","Type":"ContainerStarted","Data":"2e12779937a8dc09658f66df6719cbaaef50cb3ba21887684b6a4b7b977cabea"} Dec 02 03:08:24 crc kubenswrapper[4884]: I1202 03:08:24.118514 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-h4s98/crc-debug-c9hgj" podStartSLOduration=1.118492423 podStartE2EDuration="1.118492423s" podCreationTimestamp="2025-12-02 03:08:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-02 03:08:24.110014076 +0000 UTC m=+5520.785850960" watchObservedRunningTime="2025-12-02 03:08:24.118492423 +0000 UTC m=+5520.794329307" Dec 02 03:08:25 crc kubenswrapper[4884]: I1202 03:08:25.108121 4884 generic.go:334] "Generic (PLEG): container finished" podID="973d6331-20f4-4cce-a5c2-163e11f76d22" containerID="1032ec5163c2492db9cacf75bd0ea2eec65d5903dfa4b08fdb4d2ceb746c466f" exitCode=0 Dec 02 03:08:25 crc kubenswrapper[4884]: I1202 03:08:25.108333 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h4s98/crc-debug-c9hgj" event={"ID":"973d6331-20f4-4cce-a5c2-163e11f76d22","Type":"ContainerDied","Data":"1032ec5163c2492db9cacf75bd0ea2eec65d5903dfa4b08fdb4d2ceb746c466f"} Dec 02 03:08:26 crc kubenswrapper[4884]: I1202 03:08:26.205803 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/crc-debug-c9hgj" Dec 02 03:08:26 crc kubenswrapper[4884]: I1202 03:08:26.270209 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-h4s98/crc-debug-c9hgj"] Dec 02 03:08:26 crc kubenswrapper[4884]: I1202 03:08:26.279171 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-h4s98/crc-debug-c9hgj"] Dec 02 03:08:26 crc kubenswrapper[4884]: I1202 03:08:26.299316 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcpcb\" (UniqueName: \"kubernetes.io/projected/973d6331-20f4-4cce-a5c2-163e11f76d22-kube-api-access-bcpcb\") pod \"973d6331-20f4-4cce-a5c2-163e11f76d22\" (UID: \"973d6331-20f4-4cce-a5c2-163e11f76d22\") " Dec 02 03:08:26 crc kubenswrapper[4884]: I1202 03:08:26.299599 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/973d6331-20f4-4cce-a5c2-163e11f76d22-host\") pod \"973d6331-20f4-4cce-a5c2-163e11f76d22\" (UID: \"973d6331-20f4-4cce-a5c2-163e11f76d22\") " Dec 02 03:08:26 crc kubenswrapper[4884]: I1202 03:08:26.299645 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/973d6331-20f4-4cce-a5c2-163e11f76d22-host" (OuterVolumeSpecName: "host") pod "973d6331-20f4-4cce-a5c2-163e11f76d22" (UID: "973d6331-20f4-4cce-a5c2-163e11f76d22"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 03:08:26 crc kubenswrapper[4884]: I1202 03:08:26.300156 4884 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/973d6331-20f4-4cce-a5c2-163e11f76d22-host\") on node \"crc\" DevicePath \"\"" Dec 02 03:08:26 crc kubenswrapper[4884]: I1202 03:08:26.304299 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/973d6331-20f4-4cce-a5c2-163e11f76d22-kube-api-access-bcpcb" (OuterVolumeSpecName: "kube-api-access-bcpcb") pod "973d6331-20f4-4cce-a5c2-163e11f76d22" (UID: "973d6331-20f4-4cce-a5c2-163e11f76d22"). InnerVolumeSpecName "kube-api-access-bcpcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 03:08:26 crc kubenswrapper[4884]: I1202 03:08:26.401970 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcpcb\" (UniqueName: \"kubernetes.io/projected/973d6331-20f4-4cce-a5c2-163e11f76d22-kube-api-access-bcpcb\") on node \"crc\" DevicePath \"\"" Dec 02 03:08:27 crc kubenswrapper[4884]: I1202 03:08:27.124556 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e12779937a8dc09658f66df6719cbaaef50cb3ba21887684b6a4b7b977cabea" Dec 02 03:08:27 crc kubenswrapper[4884]: I1202 03:08:27.124600 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/crc-debug-c9hgj" Dec 02 03:08:27 crc kubenswrapper[4884]: I1202 03:08:27.492649 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h4s98/crc-debug-pd2rg"] Dec 02 03:08:27 crc kubenswrapper[4884]: E1202 03:08:27.493349 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="973d6331-20f4-4cce-a5c2-163e11f76d22" containerName="container-00" Dec 02 03:08:27 crc kubenswrapper[4884]: I1202 03:08:27.493362 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="973d6331-20f4-4cce-a5c2-163e11f76d22" containerName="container-00" Dec 02 03:08:27 crc kubenswrapper[4884]: I1202 03:08:27.493677 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="973d6331-20f4-4cce-a5c2-163e11f76d22" containerName="container-00" Dec 02 03:08:27 crc kubenswrapper[4884]: I1202 03:08:27.494339 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/crc-debug-pd2rg" Dec 02 03:08:27 crc kubenswrapper[4884]: I1202 03:08:27.627040 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="973d6331-20f4-4cce-a5c2-163e11f76d22" path="/var/lib/kubelet/pods/973d6331-20f4-4cce-a5c2-163e11f76d22/volumes" Dec 02 03:08:27 crc kubenswrapper[4884]: I1202 03:08:27.649187 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wlk9\" (UniqueName: \"kubernetes.io/projected/eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d-kube-api-access-2wlk9\") pod \"crc-debug-pd2rg\" (UID: \"eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d\") " pod="openshift-must-gather-h4s98/crc-debug-pd2rg" Dec 02 03:08:27 crc kubenswrapper[4884]: I1202 03:08:27.649406 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d-host\") pod \"crc-debug-pd2rg\" (UID: \"eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d\") " pod="openshift-must-gather-h4s98/crc-debug-pd2rg" Dec 02 03:08:27 crc kubenswrapper[4884]: I1202 03:08:27.751315 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d-host\") pod \"crc-debug-pd2rg\" (UID: \"eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d\") " pod="openshift-must-gather-h4s98/crc-debug-pd2rg" Dec 02 03:08:27 crc kubenswrapper[4884]: I1202 03:08:27.751410 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wlk9\" (UniqueName: \"kubernetes.io/projected/eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d-kube-api-access-2wlk9\") pod \"crc-debug-pd2rg\" (UID: \"eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d\") " pod="openshift-must-gather-h4s98/crc-debug-pd2rg" Dec 02 03:08:27 crc kubenswrapper[4884]: I1202 03:08:27.751471 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d-host\") pod \"crc-debug-pd2rg\" (UID: \"eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d\") " pod="openshift-must-gather-h4s98/crc-debug-pd2rg" Dec 02 03:08:27 crc kubenswrapper[4884]: I1202 03:08:27.774784 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wlk9\" (UniqueName: \"kubernetes.io/projected/eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d-kube-api-access-2wlk9\") pod \"crc-debug-pd2rg\" (UID: \"eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d\") " pod="openshift-must-gather-h4s98/crc-debug-pd2rg" Dec 02 03:08:27 crc kubenswrapper[4884]: I1202 03:08:27.823383 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/crc-debug-pd2rg" Dec 02 03:08:28 crc kubenswrapper[4884]: I1202 03:08:28.137399 4884 generic.go:334] "Generic (PLEG): container finished" podID="eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d" containerID="ec6f091ff120b2f79b1eb8b98fe0d1cd4bec0cb63d43e6330a60dd59a48ed7c8" exitCode=0 Dec 02 03:08:28 crc kubenswrapper[4884]: I1202 03:08:28.137509 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h4s98/crc-debug-pd2rg" event={"ID":"eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d","Type":"ContainerDied","Data":"ec6f091ff120b2f79b1eb8b98fe0d1cd4bec0cb63d43e6330a60dd59a48ed7c8"} Dec 02 03:08:28 crc kubenswrapper[4884]: I1202 03:08:28.137715 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h4s98/crc-debug-pd2rg" event={"ID":"eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d","Type":"ContainerStarted","Data":"39e31570df44073fe163ef66bf156865d00ed6b02293ecf70e261e8e4697bdba"} Dec 02 03:08:28 crc kubenswrapper[4884]: I1202 03:08:28.176774 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-h4s98/crc-debug-pd2rg"] Dec 02 03:08:28 crc kubenswrapper[4884]: I1202 03:08:28.189629 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-h4s98/crc-debug-pd2rg"] Dec 02 03:08:29 crc kubenswrapper[4884]: I1202 03:08:29.232682 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/crc-debug-pd2rg" Dec 02 03:08:29 crc kubenswrapper[4884]: I1202 03:08:29.382974 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d-host\") pod \"eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d\" (UID: \"eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d\") " Dec 02 03:08:29 crc kubenswrapper[4884]: I1202 03:08:29.383171 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wlk9\" (UniqueName: \"kubernetes.io/projected/eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d-kube-api-access-2wlk9\") pod \"eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d\" (UID: \"eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d\") " Dec 02 03:08:29 crc kubenswrapper[4884]: I1202 03:08:29.384793 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d-host" (OuterVolumeSpecName: "host") pod "eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d" (UID: "eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 02 03:08:29 crc kubenswrapper[4884]: I1202 03:08:29.388521 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d-kube-api-access-2wlk9" (OuterVolumeSpecName: "kube-api-access-2wlk9") pod "eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d" (UID: "eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d"). InnerVolumeSpecName "kube-api-access-2wlk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 03:08:29 crc kubenswrapper[4884]: I1202 03:08:29.485836 4884 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d-host\") on node \"crc\" DevicePath \"\"" Dec 02 03:08:29 crc kubenswrapper[4884]: I1202 03:08:29.485892 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wlk9\" (UniqueName: \"kubernetes.io/projected/eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d-kube-api-access-2wlk9\") on node \"crc\" DevicePath \"\"" Dec 02 03:08:29 crc kubenswrapper[4884]: I1202 03:08:29.630003 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d" path="/var/lib/kubelet/pods/eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d/volumes" Dec 02 03:08:30 crc kubenswrapper[4884]: I1202 03:08:30.156414 4884 scope.go:117] "RemoveContainer" containerID="ec6f091ff120b2f79b1eb8b98fe0d1cd4bec0cb63d43e6330a60dd59a48ed7c8" Dec 02 03:08:30 crc kubenswrapper[4884]: I1202 03:08:30.156458 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/crc-debug-pd2rg" Dec 02 03:08:36 crc kubenswrapper[4884]: I1202 03:08:36.615707 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:08:36 crc kubenswrapper[4884]: E1202 03:08:36.616849 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:08:50 crc kubenswrapper[4884]: I1202 03:08:50.616263 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:08:50 crc kubenswrapper[4884]: E1202 03:08:50.617367 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:08:58 crc kubenswrapper[4884]: I1202 03:08:58.147087 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5fb984d75d-76xfp_6059a81d-e008-4cba-8af1-9b093989ddf2/barbican-api-log/0.log" Dec 02 03:08:58 crc kubenswrapper[4884]: I1202 03:08:58.151665 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5fb984d75d-76xfp_6059a81d-e008-4cba-8af1-9b093989ddf2/barbican-api/0.log" Dec 02 03:08:58 crc kubenswrapper[4884]: I1202 03:08:58.311103 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-57b6695dcb-8c27t_e0163d38-d25f-460e-b563-7d7198b52999/barbican-keystone-listener/0.log" Dec 02 03:08:58 crc kubenswrapper[4884]: I1202 03:08:58.377202 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-57b6695dcb-8c27t_e0163d38-d25f-460e-b563-7d7198b52999/barbican-keystone-listener-log/0.log" Dec 02 03:08:58 crc kubenswrapper[4884]: I1202 03:08:58.436246 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7b4498cc99-xx9bb_80449fc3-b0b7-45fd-9027-49ca53a34bb6/barbican-worker/0.log" Dec 02 03:08:58 crc kubenswrapper[4884]: I1202 03:08:58.505548 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7b4498cc99-xx9bb_80449fc3-b0b7-45fd-9027-49ca53a34bb6/barbican-worker-log/0.log" Dec 02 03:08:58 crc kubenswrapper[4884]: I1202 03:08:58.665369 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-5v28p_49a3a2e3-b9ff-49bb-b850-5a516270f73b/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:08:58 crc kubenswrapper[4884]: I1202 03:08:58.748102 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0bfa748c-a54a-4a8c-bae8-f87fefe8a81a/ceilometer-central-agent/0.log" Dec 02 03:08:58 crc kubenswrapper[4884]: I1202 03:08:58.783105 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0bfa748c-a54a-4a8c-bae8-f87fefe8a81a/ceilometer-notification-agent/0.log" Dec 02 03:08:58 crc kubenswrapper[4884]: I1202 03:08:58.862962 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0bfa748c-a54a-4a8c-bae8-f87fefe8a81a/sg-core/0.log" Dec 02 03:08:58 crc kubenswrapper[4884]: I1202 03:08:58.885630 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0bfa748c-a54a-4a8c-bae8-f87fefe8a81a/proxy-httpd/0.log" Dec 02 03:08:59 crc kubenswrapper[4884]: I1202 03:08:59.054840 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e174ee23-a3a9-4d96-a76f-d63e984698d2/cinder-api/0.log" Dec 02 03:08:59 crc kubenswrapper[4884]: I1202 03:08:59.081327 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e174ee23-a3a9-4d96-a76f-d63e984698d2/cinder-api-log/0.log" Dec 02 03:08:59 crc kubenswrapper[4884]: I1202 03:08:59.162248 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4/cinder-scheduler/0.log" Dec 02 03:08:59 crc kubenswrapper[4884]: I1202 03:08:59.291725 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_54e1660b-e7e3-4e25-8e0c-aebc9a01c8f4/probe/0.log" Dec 02 03:08:59 crc kubenswrapper[4884]: I1202 03:08:59.305615 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4gns2_c129a2ce-c6ab-47b5-af8f-55dea1e97d90/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:08:59 crc kubenswrapper[4884]: I1202 03:08:59.483446 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6cd9bffc9-zqncs_0bfdff8a-051f-4d73-94c8-9472af62a278/init/0.log" Dec 02 03:08:59 crc kubenswrapper[4884]: I1202 03:08:59.494721 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-tlflf_e40740f3-80d8-46da-bfdc-467f13f8be12/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:08:59 crc kubenswrapper[4884]: I1202 03:08:59.660566 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6cd9bffc9-zqncs_0bfdff8a-051f-4d73-94c8-9472af62a278/init/0.log" Dec 02 03:08:59 crc kubenswrapper[4884]: I1202 03:08:59.729837 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-bzw8m_c5829ac7-a36e-4c26-a03b-d88518e1361f/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:08:59 crc kubenswrapper[4884]: I1202 03:08:59.813951 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6cd9bffc9-zqncs_0bfdff8a-051f-4d73-94c8-9472af62a278/dnsmasq-dns/0.log" Dec 02 03:08:59 crc kubenswrapper[4884]: I1202 03:08:59.929090 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_18baf695-0c9e-4c27-a116-13257bd7d281/glance-httpd/0.log" Dec 02 03:08:59 crc kubenswrapper[4884]: I1202 03:08:59.980164 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_18baf695-0c9e-4c27-a116-13257bd7d281/glance-log/0.log" Dec 02 03:09:00 crc kubenswrapper[4884]: I1202 03:09:00.119380 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c/glance-log/0.log" Dec 02 03:09:00 crc kubenswrapper[4884]: I1202 03:09:00.159063 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d09bd0e7-6fdb-45f3-a2a9-c20fbc1b386c/glance-httpd/0.log" Dec 02 03:09:00 crc kubenswrapper[4884]: I1202 03:09:00.315469 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-8687f948dd-5zg8m_19afec6c-37bd-4e1a-9f96-c7e37c1e7748/horizon/0.log" Dec 02 03:09:00 crc kubenswrapper[4884]: I1202 03:09:00.519784 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-trsmt_32e5e682-f72d-44bd-93c1-0fc27c02f99a/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:09:00 crc kubenswrapper[4884]: I1202 03:09:00.697069 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-mqrmf_860075b8-af90-4ffa-90aa-1c13451eb458/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:09:00 crc kubenswrapper[4884]: I1202 03:09:00.962201 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-8687f948dd-5zg8m_19afec6c-37bd-4e1a-9f96-c7e37c1e7748/horizon-log/0.log" Dec 02 03:09:00 crc kubenswrapper[4884]: I1202 03:09:00.968417 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29410681-v4hzm_8fdf7bbf-bbc0-4280-99b9-8c87c8f6635c/keystone-cron/0.log" Dec 02 03:09:01 crc kubenswrapper[4884]: I1202 03:09:01.156374 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29410741-ngbgf_9123c6f1-236e-4598-8309-9f6f3784613c/keystone-cron/0.log" Dec 02 03:09:01 crc kubenswrapper[4884]: I1202 03:09:01.157246 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-87844dc44-b7fcj_8fad23d3-2fad-430e-a78f-7cb9e778f9ae/keystone-api/0.log" Dec 02 03:09:01 crc kubenswrapper[4884]: I1202 03:09:01.173265 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_9537581d-d462-4e9d-992e-15c9754c3fb3/kube-state-metrics/0.log" Dec 02 03:09:01 crc kubenswrapper[4884]: I1202 03:09:01.354358 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-nhgnb_ba6d4723-02c0-46ce-bfd0-afa42fd31d7d/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:09:01 crc kubenswrapper[4884]: I1202 03:09:01.678998 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-mmzgs_9703a9cb-2e1a-4e9f-a862-8025a775ac87/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:09:01 crc kubenswrapper[4884]: I1202 03:09:01.822503 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7b6477d857-jzgzr_18e31af0-c0a9-4dc4-8281-5eef1a4477e4/neutron-httpd/0.log" Dec 02 03:09:01 crc kubenswrapper[4884]: I1202 03:09:01.825353 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7b6477d857-jzgzr_18e31af0-c0a9-4dc4-8281-5eef1a4477e4/neutron-api/0.log" Dec 02 03:09:02 crc kubenswrapper[4884]: I1202 03:09:02.297875 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_67d11b36-d993-46f4-9ca9-813df373dcab/nova-cell0-conductor-conductor/0.log" Dec 02 03:09:02 crc kubenswrapper[4884]: I1202 03:09:02.580854 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_5993c2f9-ca6d-454f-ae88-08b91e419b3f/nova-cell1-conductor-conductor/0.log" Dec 02 03:09:02 crc kubenswrapper[4884]: I1202 03:09:02.874167 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_c72fc8df-eefd-43aa-879f-ac397417aa90/nova-cell1-novncproxy-novncproxy/0.log" Dec 02 03:09:03 crc kubenswrapper[4884]: I1202 03:09:03.017337 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d/nova-api-log/0.log" Dec 02 03:09:03 crc kubenswrapper[4884]: I1202 03:09:03.033644 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-lqsdb_7ee4fe1d-9395-4e57-8d2c-7dca61af8238/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:09:03 crc kubenswrapper[4884]: I1202 03:09:03.220222 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_0e80ddec-bfb2-420d-aeb4-0a27e56a5f3d/nova-api-api/0.log" Dec 02 03:09:03 crc kubenswrapper[4884]: I1202 03:09:03.360516 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9e9beb88-6b94-45ad-ac2b-a86abd37c944/nova-metadata-log/0.log" Dec 02 03:09:03 crc kubenswrapper[4884]: I1202 03:09:03.559876 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a8992d93-8785-4e52-af5b-73af8e0f56e5/mysql-bootstrap/0.log" Dec 02 03:09:03 crc kubenswrapper[4884]: I1202 03:09:03.688391 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1a34c2cd-4194-4086-95a2-3d016faa05db/nova-scheduler-scheduler/0.log" Dec 02 03:09:03 crc kubenswrapper[4884]: I1202 03:09:03.835437 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a8992d93-8785-4e52-af5b-73af8e0f56e5/galera/0.log" Dec 02 03:09:03 crc kubenswrapper[4884]: I1202 03:09:03.856242 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a8992d93-8785-4e52-af5b-73af8e0f56e5/mysql-bootstrap/0.log" Dec 02 03:09:04 crc kubenswrapper[4884]: I1202 03:09:04.060304 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_63a48d7f-b557-4f21-93d6-8a18e96eea16/mysql-bootstrap/0.log" Dec 02 03:09:04 crc kubenswrapper[4884]: I1202 03:09:04.263737 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_63a48d7f-b557-4f21-93d6-8a18e96eea16/galera/0.log" Dec 02 03:09:04 crc kubenswrapper[4884]: I1202 03:09:04.293561 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_63a48d7f-b557-4f21-93d6-8a18e96eea16/mysql-bootstrap/0.log" Dec 02 03:09:04 crc kubenswrapper[4884]: I1202 03:09:04.510147 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_b7e641c0-71a1-4786-aeb9-1fbba21c3e49/openstackclient/0.log" Dec 02 03:09:04 crc kubenswrapper[4884]: I1202 03:09:04.542255 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-9wkcq_cdabd8d6-7f1f-4629-b6cb-1be182bd9b66/ovn-controller/0.log" Dec 02 03:09:04 crc kubenswrapper[4884]: I1202 03:09:04.696313 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-hplrx_dc44c0f6-5201-4b7a-85e9-db2a1766c237/openstack-network-exporter/0.log" Dec 02 03:09:05 crc kubenswrapper[4884]: I1202 03:09:05.207153 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_9e9beb88-6b94-45ad-ac2b-a86abd37c944/nova-metadata-metadata/0.log" Dec 02 03:09:05 crc kubenswrapper[4884]: I1202 03:09:05.393516 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bdv77_2536c28e-5917-4fc9-b37a-2f9212574dc5/ovsdb-server-init/0.log" Dec 02 03:09:05 crc kubenswrapper[4884]: I1202 03:09:05.514960 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bdv77_2536c28e-5917-4fc9-b37a-2f9212574dc5/ovs-vswitchd/0.log" Dec 02 03:09:05 crc kubenswrapper[4884]: I1202 03:09:05.528452 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bdv77_2536c28e-5917-4fc9-b37a-2f9212574dc5/ovsdb-server-init/0.log" Dec 02 03:09:05 crc kubenswrapper[4884]: I1202 03:09:05.614455 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:09:05 crc kubenswrapper[4884]: E1202 03:09:05.614658 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:09:05 crc kubenswrapper[4884]: I1202 03:09:05.624023 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bdv77_2536c28e-5917-4fc9-b37a-2f9212574dc5/ovsdb-server/0.log" Dec 02 03:09:05 crc kubenswrapper[4884]: I1202 03:09:05.753366 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-z8nsq_bd5930b2-0f14-44df-bb7c-7d5e08072485/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:09:05 crc kubenswrapper[4884]: I1202 03:09:05.881892 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ffe8e419-32a1-46e9-b2b3-a5f90c18dcde/openstack-network-exporter/0.log" Dec 02 03:09:05 crc kubenswrapper[4884]: I1202 03:09:05.968840 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_ffe8e419-32a1-46e9-b2b3-a5f90c18dcde/ovn-northd/0.log" Dec 02 03:09:05 crc kubenswrapper[4884]: I1202 03:09:05.997969 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d16c1f3a-feb6-4809-b3d0-7306fb3d20ba/openstack-network-exporter/0.log" Dec 02 03:09:06 crc kubenswrapper[4884]: I1202 03:09:06.087808 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_d16c1f3a-feb6-4809-b3d0-7306fb3d20ba/ovsdbserver-nb/0.log" Dec 02 03:09:06 crc kubenswrapper[4884]: I1202 03:09:06.246694 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_da645b5a-4282-4b4f-8eda-8713f990e51f/ovsdbserver-sb/0.log" Dec 02 03:09:06 crc kubenswrapper[4884]: I1202 03:09:06.254108 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_da645b5a-4282-4b4f-8eda-8713f990e51f/openstack-network-exporter/0.log" Dec 02 03:09:06 crc kubenswrapper[4884]: I1202 03:09:06.561048 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-595d8586d8-j5926_1f4a5e6c-067d-4f50-b053-950fb802d52e/placement-api/0.log" Dec 02 03:09:06 crc kubenswrapper[4884]: I1202 03:09:06.613332 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_827e7eb5-0a90-4677-9d1f-50185bf48b52/init-config-reloader/0.log" Dec 02 03:09:06 crc kubenswrapper[4884]: I1202 03:09:06.687421 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-595d8586d8-j5926_1f4a5e6c-067d-4f50-b053-950fb802d52e/placement-log/0.log" Dec 02 03:09:07 crc kubenswrapper[4884]: I1202 03:09:07.241515 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_827e7eb5-0a90-4677-9d1f-50185bf48b52/init-config-reloader/0.log" Dec 02 03:09:07 crc kubenswrapper[4884]: I1202 03:09:07.260263 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_827e7eb5-0a90-4677-9d1f-50185bf48b52/thanos-sidecar/0.log" Dec 02 03:09:07 crc kubenswrapper[4884]: I1202 03:09:07.284054 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_827e7eb5-0a90-4677-9d1f-50185bf48b52/config-reloader/0.log" Dec 02 03:09:07 crc kubenswrapper[4884]: I1202 03:09:07.299165 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_827e7eb5-0a90-4677-9d1f-50185bf48b52/prometheus/0.log" Dec 02 03:09:07 crc kubenswrapper[4884]: I1202 03:09:07.521472 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_962f0d62-e23b-437e-ac98-aea141142a36/setup-container/0.log" Dec 02 03:09:07 crc kubenswrapper[4884]: I1202 03:09:07.660633 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_962f0d62-e23b-437e-ac98-aea141142a36/setup-container/0.log" Dec 02 03:09:07 crc kubenswrapper[4884]: I1202 03:09:07.688498 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_962f0d62-e23b-437e-ac98-aea141142a36/rabbitmq/0.log" Dec 02 03:09:07 crc kubenswrapper[4884]: I1202 03:09:07.772346 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c/setup-container/0.log" Dec 02 03:09:08 crc kubenswrapper[4884]: I1202 03:09:08.023684 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-trskf_d2510824-0c28-4011-a19a-39c0dfb0eafd/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:09:08 crc kubenswrapper[4884]: I1202 03:09:08.035020 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c/rabbitmq/0.log" Dec 02 03:09:08 crc kubenswrapper[4884]: I1202 03:09:08.037579 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8ed44e1a-89f9-4e26-9a9c-f83ee9b0f50c/setup-container/0.log" Dec 02 03:09:08 crc kubenswrapper[4884]: I1202 03:09:08.263545 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-ql9sv_b074036a-792d-4270-af51-d3ae2dd92bda/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:09:08 crc kubenswrapper[4884]: I1202 03:09:08.300693 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-tdk62_06fb15ba-f86b-4962-8b6b-52367ad24e09/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:09:08 crc kubenswrapper[4884]: I1202 03:09:08.483217 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-4v9qb_d52773c0-4f07-46c6-9774-5fba955697be/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:09:08 crc kubenswrapper[4884]: I1202 03:09:08.580061 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-h5w4t_bedee34a-ff45-4905-9e6d-7408127cbd22/ssh-known-hosts-edpm-deployment/0.log" Dec 02 03:09:08 crc kubenswrapper[4884]: I1202 03:09:08.851950 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-558bcd5597-pglnv_c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c/proxy-server/0.log" Dec 02 03:09:08 crc kubenswrapper[4884]: I1202 03:09:08.970259 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-l2j2h_151b4ede-9694-42d6-8051-6309de26a4e4/swift-ring-rebalance/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.048110 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-558bcd5597-pglnv_c80abbe4-20e4-4a36-b25a-9e79ccdb0d0c/proxy-httpd/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.185585 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8dc2a1f5-a7a6-4ce5-a711-77455efe2224/account-auditor/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.195557 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8dc2a1f5-a7a6-4ce5-a711-77455efe2224/account-reaper/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.270652 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8dc2a1f5-a7a6-4ce5-a711-77455efe2224/account-replicator/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.327618 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8dc2a1f5-a7a6-4ce5-a711-77455efe2224/account-server/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.392613 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_dc21a47b-5514-4bd1-b164-05eb04fe2e31/memcached/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.410259 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8dc2a1f5-a7a6-4ce5-a711-77455efe2224/container-auditor/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.432505 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8dc2a1f5-a7a6-4ce5-a711-77455efe2224/container-replicator/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.513450 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8dc2a1f5-a7a6-4ce5-a711-77455efe2224/container-server/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.545838 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8dc2a1f5-a7a6-4ce5-a711-77455efe2224/container-updater/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.615738 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8dc2a1f5-a7a6-4ce5-a711-77455efe2224/object-auditor/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.652733 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8dc2a1f5-a7a6-4ce5-a711-77455efe2224/object-expirer/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.680441 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8dc2a1f5-a7a6-4ce5-a711-77455efe2224/object-replicator/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.742214 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8dc2a1f5-a7a6-4ce5-a711-77455efe2224/object-server/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.758166 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8dc2a1f5-a7a6-4ce5-a711-77455efe2224/object-updater/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.810973 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8dc2a1f5-a7a6-4ce5-a711-77455efe2224/rsync/0.log" Dec 02 03:09:09 crc kubenswrapper[4884]: I1202 03:09:09.846427 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_8dc2a1f5-a7a6-4ce5-a711-77455efe2224/swift-recon-cron/0.log" Dec 02 03:09:10 crc kubenswrapper[4884]: I1202 03:09:10.002208 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-wl5xz_7098a534-9f0a-457d-a3d2-09249bd490ee/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:09:10 crc kubenswrapper[4884]: I1202 03:09:10.111630 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_56afebdd-31ad-4882-b405-53ada83cbaea/tempest-tests-tempest-tests-runner/0.log" Dec 02 03:09:10 crc kubenswrapper[4884]: I1202 03:09:10.143719 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_6cd1aa1b-940f-40ab-9776-f74981c420df/test-operator-logs-container/0.log" Dec 02 03:09:10 crc kubenswrapper[4884]: I1202 03:09:10.212814 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-nrlnq_c70940a8-abc7-4e98-bf89-378b021e5552/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 02 03:09:10 crc kubenswrapper[4884]: I1202 03:09:10.858435 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-applier-0_25acc2a6-f07c-44c1-aba6-b5f2e38d39ce/watcher-applier/0.log" Dec 02 03:09:11 crc kubenswrapper[4884]: I1202 03:09:11.131030 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_a84d6c24-16a7-43a2-a303-5befb347c8f4/watcher-api-log/0.log" Dec 02 03:09:11 crc kubenswrapper[4884]: I1202 03:09:11.586488 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-decision-engine-0_40a8993e-7365-473f-9dbd-1f34c001714b/watcher-decision-engine/0.log" Dec 02 03:09:13 crc kubenswrapper[4884]: I1202 03:09:13.203935 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_watcher-api-0_a84d6c24-16a7-43a2-a303-5befb347c8f4/watcher-api/0.log" Dec 02 03:09:16 crc kubenswrapper[4884]: I1202 03:09:16.614262 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:09:16 crc kubenswrapper[4884]: E1202 03:09:16.615280 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:09:18 crc kubenswrapper[4884]: I1202 03:09:18.749895 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="63a48d7f-b557-4f21-93d6-8a18e96eea16" containerName="galera" probeResult="failure" output="command timed out" Dec 02 03:09:27 crc kubenswrapper[4884]: I1202 03:09:27.614853 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:09:28 crc kubenswrapper[4884]: I1202 03:09:28.806184 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"783609d4dac9b374973e24fed69e04030080c197a1b2e0ea5d35a59a53de30f9"} Dec 02 03:09:32 crc kubenswrapper[4884]: I1202 03:09:32.825621 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2cb4f"] Dec 02 03:09:32 crc kubenswrapper[4884]: E1202 03:09:32.826982 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d" containerName="container-00" Dec 02 03:09:32 crc kubenswrapper[4884]: I1202 03:09:32.827004 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d" containerName="container-00" Dec 02 03:09:32 crc kubenswrapper[4884]: I1202 03:09:32.827308 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf9acb1-4d0f-44b2-a7d7-ac2fcb78f97d" containerName="container-00" Dec 02 03:09:32 crc kubenswrapper[4884]: I1202 03:09:32.829810 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:32 crc kubenswrapper[4884]: I1202 03:09:32.854191 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2cb4f"] Dec 02 03:09:32 crc kubenswrapper[4884]: I1202 03:09:32.921903 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7q5l\" (UniqueName: \"kubernetes.io/projected/ef66ab1f-432e-4274-a31b-68e8690c6f00-kube-api-access-x7q5l\") pod \"redhat-marketplace-2cb4f\" (UID: \"ef66ab1f-432e-4274-a31b-68e8690c6f00\") " pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:32 crc kubenswrapper[4884]: I1202 03:09:32.922130 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef66ab1f-432e-4274-a31b-68e8690c6f00-utilities\") pod \"redhat-marketplace-2cb4f\" (UID: \"ef66ab1f-432e-4274-a31b-68e8690c6f00\") " pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:32 crc kubenswrapper[4884]: I1202 03:09:32.922195 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef66ab1f-432e-4274-a31b-68e8690c6f00-catalog-content\") pod \"redhat-marketplace-2cb4f\" (UID: \"ef66ab1f-432e-4274-a31b-68e8690c6f00\") " pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:33 crc kubenswrapper[4884]: I1202 03:09:33.036278 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7q5l\" (UniqueName: \"kubernetes.io/projected/ef66ab1f-432e-4274-a31b-68e8690c6f00-kube-api-access-x7q5l\") pod \"redhat-marketplace-2cb4f\" (UID: \"ef66ab1f-432e-4274-a31b-68e8690c6f00\") " pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:33 crc kubenswrapper[4884]: I1202 03:09:33.036404 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef66ab1f-432e-4274-a31b-68e8690c6f00-utilities\") pod \"redhat-marketplace-2cb4f\" (UID: \"ef66ab1f-432e-4274-a31b-68e8690c6f00\") " pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:33 crc kubenswrapper[4884]: I1202 03:09:33.036435 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef66ab1f-432e-4274-a31b-68e8690c6f00-catalog-content\") pod \"redhat-marketplace-2cb4f\" (UID: \"ef66ab1f-432e-4274-a31b-68e8690c6f00\") " pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:33 crc kubenswrapper[4884]: I1202 03:09:33.036935 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef66ab1f-432e-4274-a31b-68e8690c6f00-catalog-content\") pod \"redhat-marketplace-2cb4f\" (UID: \"ef66ab1f-432e-4274-a31b-68e8690c6f00\") " pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:33 crc kubenswrapper[4884]: I1202 03:09:33.037510 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef66ab1f-432e-4274-a31b-68e8690c6f00-utilities\") pod \"redhat-marketplace-2cb4f\" (UID: \"ef66ab1f-432e-4274-a31b-68e8690c6f00\") " pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:33 crc kubenswrapper[4884]: I1202 03:09:33.065452 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7q5l\" (UniqueName: \"kubernetes.io/projected/ef66ab1f-432e-4274-a31b-68e8690c6f00-kube-api-access-x7q5l\") pod \"redhat-marketplace-2cb4f\" (UID: \"ef66ab1f-432e-4274-a31b-68e8690c6f00\") " pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:33 crc kubenswrapper[4884]: I1202 03:09:33.169563 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:33 crc kubenswrapper[4884]: I1202 03:09:33.677183 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2cb4f"] Dec 02 03:09:33 crc kubenswrapper[4884]: I1202 03:09:33.873471 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2cb4f" event={"ID":"ef66ab1f-432e-4274-a31b-68e8690c6f00","Type":"ContainerStarted","Data":"41b8f5e6e95901eefbf3d9ad63680cb50ca5b4496c7ff450fb49769022c6ee10"} Dec 02 03:09:34 crc kubenswrapper[4884]: I1202 03:09:34.893052 4884 generic.go:334] "Generic (PLEG): container finished" podID="ef66ab1f-432e-4274-a31b-68e8690c6f00" containerID="718b60f5f8bf8d6cc02751a25b7054cf5ff59c44cb1564aecb31495dd4c65727" exitCode=0 Dec 02 03:09:34 crc kubenswrapper[4884]: I1202 03:09:34.893156 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2cb4f" event={"ID":"ef66ab1f-432e-4274-a31b-68e8690c6f00","Type":"ContainerDied","Data":"718b60f5f8bf8d6cc02751a25b7054cf5ff59c44cb1564aecb31495dd4c65727"} Dec 02 03:09:36 crc kubenswrapper[4884]: I1202 03:09:36.959315 4884 generic.go:334] "Generic (PLEG): container finished" podID="ef66ab1f-432e-4274-a31b-68e8690c6f00" containerID="2b962ef33a9791ef43af9d9bbff7e4b157a6c6f3de8a51852e259399f71d66d7" exitCode=0 Dec 02 03:09:36 crc kubenswrapper[4884]: I1202 03:09:36.959388 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2cb4f" event={"ID":"ef66ab1f-432e-4274-a31b-68e8690c6f00","Type":"ContainerDied","Data":"2b962ef33a9791ef43af9d9bbff7e4b157a6c6f3de8a51852e259399f71d66d7"} Dec 02 03:09:37 crc kubenswrapper[4884]: I1202 03:09:37.974195 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2cb4f" event={"ID":"ef66ab1f-432e-4274-a31b-68e8690c6f00","Type":"ContainerStarted","Data":"e9822b1e0d3d2da8a4e8ca7a2775a0b81fad15877d18acf4f6666c01dbb7357f"} Dec 02 03:09:37 crc kubenswrapper[4884]: I1202 03:09:37.996334 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2cb4f" podStartSLOduration=3.377131244 podStartE2EDuration="5.996302032s" podCreationTimestamp="2025-12-02 03:09:32 +0000 UTC" firstStartedPulling="2025-12-02 03:09:34.89798187 +0000 UTC m=+5591.573818794" lastFinishedPulling="2025-12-02 03:09:37.517152688 +0000 UTC m=+5594.192989582" observedRunningTime="2025-12-02 03:09:37.99416923 +0000 UTC m=+5594.670006124" watchObservedRunningTime="2025-12-02 03:09:37.996302032 +0000 UTC m=+5594.672138926" Dec 02 03:09:38 crc kubenswrapper[4884]: I1202 03:09:38.760890 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs_0dc571b2-e65a-41d5-b47c-8acbb09cbe28/util/0.log" Dec 02 03:09:38 crc kubenswrapper[4884]: I1202 03:09:38.923935 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs_0dc571b2-e65a-41d5-b47c-8acbb09cbe28/pull/0.log" Dec 02 03:09:38 crc kubenswrapper[4884]: I1202 03:09:38.925242 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs_0dc571b2-e65a-41d5-b47c-8acbb09cbe28/util/0.log" Dec 02 03:09:38 crc kubenswrapper[4884]: I1202 03:09:38.933446 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs_0dc571b2-e65a-41d5-b47c-8acbb09cbe28/pull/0.log" Dec 02 03:09:39 crc kubenswrapper[4884]: I1202 03:09:39.113019 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs_0dc571b2-e65a-41d5-b47c-8acbb09cbe28/pull/0.log" Dec 02 03:09:39 crc kubenswrapper[4884]: I1202 03:09:39.116020 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs_0dc571b2-e65a-41d5-b47c-8acbb09cbe28/extract/0.log" Dec 02 03:09:39 crc kubenswrapper[4884]: I1202 03:09:39.135235 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_290026b6971964c4d0a1d2e683c792af6383e2eb1eef0901c02d82c5b2q2xfs_0dc571b2-e65a-41d5-b47c-8acbb09cbe28/util/0.log" Dec 02 03:09:39 crc kubenswrapper[4884]: I1202 03:09:39.282517 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-fspk2_50955276-266a-4e3d-9537-0efccdd1028d/kube-rbac-proxy/0.log" Dec 02 03:09:39 crc kubenswrapper[4884]: I1202 03:09:39.348407 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-fspk2_50955276-266a-4e3d-9537-0efccdd1028d/manager/0.log" Dec 02 03:09:39 crc kubenswrapper[4884]: I1202 03:09:39.351374 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-7tjcj_6c738783-6d27-48e8-8f8e-a34f0f0ecef8/kube-rbac-proxy/0.log" Dec 02 03:09:39 crc kubenswrapper[4884]: I1202 03:09:39.515196 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-7tjcj_6c738783-6d27-48e8-8f8e-a34f0f0ecef8/manager/0.log" Dec 02 03:09:39 crc kubenswrapper[4884]: I1202 03:09:39.522455 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-h8944_ccee2404-f917-4dc3-bd71-be8c8aaf5362/kube-rbac-proxy/0.log" Dec 02 03:09:39 crc kubenswrapper[4884]: I1202 03:09:39.526988 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-h8944_ccee2404-f917-4dc3-bd71-be8c8aaf5362/manager/0.log" Dec 02 03:09:39 crc kubenswrapper[4884]: I1202 03:09:39.685043 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-qpg2b_15629082-5a45-439e-b98e-57083adea72c/kube-rbac-proxy/0.log" Dec 02 03:09:39 crc kubenswrapper[4884]: I1202 03:09:39.762741 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-668d9c48b9-qpg2b_15629082-5a45-439e-b98e-57083adea72c/manager/0.log" Dec 02 03:09:39 crc kubenswrapper[4884]: I1202 03:09:39.878272 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4pvr8_68a5e45a-0241-4dad-97d5-3855e1128b01/kube-rbac-proxy/0.log" Dec 02 03:09:39 crc kubenswrapper[4884]: I1202 03:09:39.918539 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-4pvr8_68a5e45a-0241-4dad-97d5-3855e1128b01/manager/0.log" Dec 02 03:09:39 crc kubenswrapper[4884]: I1202 03:09:39.946897 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-j49bt_4cbce330-7938-453f-989a-c29213cf9118/kube-rbac-proxy/0.log" Dec 02 03:09:40 crc kubenswrapper[4884]: I1202 03:09:40.479015 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-j49bt_4cbce330-7938-453f-989a-c29213cf9118/manager/0.log" Dec 02 03:09:40 crc kubenswrapper[4884]: I1202 03:09:40.482354 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-nhlqg_6770b6e2-f0dc-4a64-96c7-3d0854fadf3d/kube-rbac-proxy/0.log" Dec 02 03:09:40 crc kubenswrapper[4884]: I1202 03:09:40.635043 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-nhlqg_6770b6e2-f0dc-4a64-96c7-3d0854fadf3d/manager/0.log" Dec 02 03:09:40 crc kubenswrapper[4884]: I1202 03:09:40.664934 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-kszl7_f3bc5025-5eac-4f60-a65c-c1abda8e6367/kube-rbac-proxy/0.log" Dec 02 03:09:40 crc kubenswrapper[4884]: I1202 03:09:40.686806 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-kszl7_f3bc5025-5eac-4f60-a65c-c1abda8e6367/manager/0.log" Dec 02 03:09:40 crc kubenswrapper[4884]: I1202 03:09:40.835192 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-kmj5d_d7fbdcb3-5250-4b03-b182-47506f9e0c50/kube-rbac-proxy/0.log" Dec 02 03:09:40 crc kubenswrapper[4884]: I1202 03:09:40.910449 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-546d4bdf48-kmj5d_d7fbdcb3-5250-4b03-b182-47506f9e0c50/manager/0.log" Dec 02 03:09:40 crc kubenswrapper[4884]: I1202 03:09:40.975317 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-dm29n_8b6aa6b7-93b2-4ba1-9a01-10c47d21df36/kube-rbac-proxy/0.log" Dec 02 03:09:41 crc kubenswrapper[4884]: I1202 03:09:41.037961 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6546668bfd-dm29n_8b6aa6b7-93b2-4ba1-9a01-10c47d21df36/manager/0.log" Dec 02 03:09:41 crc kubenswrapper[4884]: I1202 03:09:41.106632 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-d2df5_17378750-b1cf-4515-81f4-638404739f0f/kube-rbac-proxy/0.log" Dec 02 03:09:41 crc kubenswrapper[4884]: I1202 03:09:41.185210 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-d2df5_17378750-b1cf-4515-81f4-638404739f0f/manager/0.log" Dec 02 03:09:41 crc kubenswrapper[4884]: I1202 03:09:41.258903 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-r8hgf_8e4db64b-9154-48cd-bd3c-463038b4dd93/kube-rbac-proxy/0.log" Dec 02 03:09:41 crc kubenswrapper[4884]: I1202 03:09:41.352876 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-r8hgf_8e4db64b-9154-48cd-bd3c-463038b4dd93/manager/0.log" Dec 02 03:09:41 crc kubenswrapper[4884]: I1202 03:09:41.952076 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-vz95x_58dc93d7-c6f5-4927-9a80-887c561cbe3b/kube-rbac-proxy/0.log" Dec 02 03:09:41 crc kubenswrapper[4884]: I1202 03:09:41.963065 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-w5xb8_3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f/kube-rbac-proxy/0.log" Dec 02 03:09:42 crc kubenswrapper[4884]: I1202 03:09:42.036032 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-w5xb8_3b3fcf6f-2aeb-47e3-bc00-1e8fdfd64a5f/manager/0.log" Dec 02 03:09:42 crc kubenswrapper[4884]: I1202 03:09:42.073801 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-vz95x_58dc93d7-c6f5-4927-9a80-887c561cbe3b/manager/0.log" Dec 02 03:09:42 crc kubenswrapper[4884]: I1202 03:09:42.325310 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd456f66_2fca9668-22fe-4ea5-a438-f72c6e9c5294/manager/0.log" Dec 02 03:09:42 crc kubenswrapper[4884]: I1202 03:09:42.328773 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd456f66_2fca9668-22fe-4ea5-a438-f72c6e9c5294/kube-rbac-proxy/0.log" Dec 02 03:09:42 crc kubenswrapper[4884]: I1202 03:09:42.612471 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-rh4qh_6eb2895e-8fa8-4e8d-b171-0c47bbccf248/registry-server/0.log" Dec 02 03:09:42 crc kubenswrapper[4884]: I1202 03:09:42.667267 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-557b87d65-9tj59_4f2e27ae-6dcb-4d32-b65b-f7d96a026d25/operator/0.log" Dec 02 03:09:42 crc kubenswrapper[4884]: I1202 03:09:42.776847 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-kf879_c9178c86-b9ac-4d5b-8145-0cc5f0e79157/kube-rbac-proxy/0.log" Dec 02 03:09:42 crc kubenswrapper[4884]: I1202 03:09:42.827194 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-kf879_c9178c86-b9ac-4d5b-8145-0cc5f0e79157/manager/0.log" Dec 02 03:09:42 crc kubenswrapper[4884]: I1202 03:09:42.921557 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-nzwwd_dd662773-0d4d-4682-9fdb-b79667483d76/kube-rbac-proxy/0.log" Dec 02 03:09:42 crc kubenswrapper[4884]: I1202 03:09:42.997169 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-nzwwd_dd662773-0d4d-4682-9fdb-b79667483d76/manager/0.log" Dec 02 03:09:43 crc kubenswrapper[4884]: I1202 03:09:43.130995 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-qjjv8_d7c0cc8c-1763-443f-a2a0-5d76d093b019/operator/0.log" Dec 02 03:09:43 crc kubenswrapper[4884]: I1202 03:09:43.169698 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:43 crc kubenswrapper[4884]: I1202 03:09:43.169758 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:43 crc kubenswrapper[4884]: I1202 03:09:43.205084 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-znmww_edf5e61d-2bf9-4eb6-9363-88e76f4e9506/kube-rbac-proxy/0.log" Dec 02 03:09:43 crc kubenswrapper[4884]: I1202 03:09:43.212225 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:43 crc kubenswrapper[4884]: I1202 03:09:43.300975 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-znmww_edf5e61d-2bf9-4eb6-9363-88e76f4e9506/manager/0.log" Dec 02 03:09:43 crc kubenswrapper[4884]: I1202 03:09:43.371094 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-x5hdr_c7c9137a-0bc9-41e0-9dc0-f5976861a3b2/kube-rbac-proxy/0.log" Dec 02 03:09:43 crc kubenswrapper[4884]: I1202 03:09:43.588906 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-nnfmw_4446fb10-1203-4063-affd-bc81cbfa4238/kube-rbac-proxy/0.log" Dec 02 03:09:43 crc kubenswrapper[4884]: I1202 03:09:43.606165 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-668dd49799-k2fql_130cad20-8dfc-4f3d-a00f-d0d5d066ef54/manager/0.log" Dec 02 03:09:43 crc kubenswrapper[4884]: I1202 03:09:43.611040 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-x5hdr_c7c9137a-0bc9-41e0-9dc0-f5976861a3b2/manager/0.log" Dec 02 03:09:43 crc kubenswrapper[4884]: I1202 03:09:43.644360 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-nnfmw_4446fb10-1203-4063-affd-bc81cbfa4238/manager/0.log" Dec 02 03:09:43 crc kubenswrapper[4884]: I1202 03:09:43.778691 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-84f44f5d76-fz96n_264f2ec8-0c0a-4fce-8ed3-eb8fafae0621/kube-rbac-proxy/0.log" Dec 02 03:09:43 crc kubenswrapper[4884]: I1202 03:09:43.821723 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-84f44f5d76-fz96n_264f2ec8-0c0a-4fce-8ed3-eb8fafae0621/manager/0.log" Dec 02 03:09:44 crc kubenswrapper[4884]: I1202 03:09:44.093971 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:44 crc kubenswrapper[4884]: I1202 03:09:44.136294 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2cb4f"] Dec 02 03:09:46 crc kubenswrapper[4884]: I1202 03:09:46.053326 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2cb4f" podUID="ef66ab1f-432e-4274-a31b-68e8690c6f00" containerName="registry-server" containerID="cri-o://e9822b1e0d3d2da8a4e8ca7a2775a0b81fad15877d18acf4f6666c01dbb7357f" gracePeriod=2 Dec 02 03:09:46 crc kubenswrapper[4884]: I1202 03:09:46.546384 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:46 crc kubenswrapper[4884]: I1202 03:09:46.702501 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7q5l\" (UniqueName: \"kubernetes.io/projected/ef66ab1f-432e-4274-a31b-68e8690c6f00-kube-api-access-x7q5l\") pod \"ef66ab1f-432e-4274-a31b-68e8690c6f00\" (UID: \"ef66ab1f-432e-4274-a31b-68e8690c6f00\") " Dec 02 03:09:46 crc kubenswrapper[4884]: I1202 03:09:46.702656 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef66ab1f-432e-4274-a31b-68e8690c6f00-catalog-content\") pod \"ef66ab1f-432e-4274-a31b-68e8690c6f00\" (UID: \"ef66ab1f-432e-4274-a31b-68e8690c6f00\") " Dec 02 03:09:46 crc kubenswrapper[4884]: I1202 03:09:46.702989 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef66ab1f-432e-4274-a31b-68e8690c6f00-utilities\") pod \"ef66ab1f-432e-4274-a31b-68e8690c6f00\" (UID: \"ef66ab1f-432e-4274-a31b-68e8690c6f00\") " Dec 02 03:09:46 crc kubenswrapper[4884]: I1202 03:09:46.703674 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef66ab1f-432e-4274-a31b-68e8690c6f00-utilities" (OuterVolumeSpecName: "utilities") pod "ef66ab1f-432e-4274-a31b-68e8690c6f00" (UID: "ef66ab1f-432e-4274-a31b-68e8690c6f00"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 03:09:46 crc kubenswrapper[4884]: I1202 03:09:46.703889 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ef66ab1f-432e-4274-a31b-68e8690c6f00-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 03:09:46 crc kubenswrapper[4884]: I1202 03:09:46.707956 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef66ab1f-432e-4274-a31b-68e8690c6f00-kube-api-access-x7q5l" (OuterVolumeSpecName: "kube-api-access-x7q5l") pod "ef66ab1f-432e-4274-a31b-68e8690c6f00" (UID: "ef66ab1f-432e-4274-a31b-68e8690c6f00"). InnerVolumeSpecName "kube-api-access-x7q5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 03:09:46 crc kubenswrapper[4884]: I1202 03:09:46.718280 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef66ab1f-432e-4274-a31b-68e8690c6f00-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ef66ab1f-432e-4274-a31b-68e8690c6f00" (UID: "ef66ab1f-432e-4274-a31b-68e8690c6f00"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 03:09:46 crc kubenswrapper[4884]: I1202 03:09:46.806028 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7q5l\" (UniqueName: \"kubernetes.io/projected/ef66ab1f-432e-4274-a31b-68e8690c6f00-kube-api-access-x7q5l\") on node \"crc\" DevicePath \"\"" Dec 02 03:09:46 crc kubenswrapper[4884]: I1202 03:09:46.806063 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ef66ab1f-432e-4274-a31b-68e8690c6f00-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.062347 4884 generic.go:334] "Generic (PLEG): container finished" podID="ef66ab1f-432e-4274-a31b-68e8690c6f00" containerID="e9822b1e0d3d2da8a4e8ca7a2775a0b81fad15877d18acf4f6666c01dbb7357f" exitCode=0 Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.062421 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2cb4f" Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.062410 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2cb4f" event={"ID":"ef66ab1f-432e-4274-a31b-68e8690c6f00","Type":"ContainerDied","Data":"e9822b1e0d3d2da8a4e8ca7a2775a0b81fad15877d18acf4f6666c01dbb7357f"} Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.062722 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2cb4f" event={"ID":"ef66ab1f-432e-4274-a31b-68e8690c6f00","Type":"ContainerDied","Data":"41b8f5e6e95901eefbf3d9ad63680cb50ca5b4496c7ff450fb49769022c6ee10"} Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.062753 4884 scope.go:117] "RemoveContainer" containerID="e9822b1e0d3d2da8a4e8ca7a2775a0b81fad15877d18acf4f6666c01dbb7357f" Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.088011 4884 scope.go:117] "RemoveContainer" containerID="2b962ef33a9791ef43af9d9bbff7e4b157a6c6f3de8a51852e259399f71d66d7" Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.090396 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2cb4f"] Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.098554 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2cb4f"] Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.106172 4884 scope.go:117] "RemoveContainer" containerID="718b60f5f8bf8d6cc02751a25b7054cf5ff59c44cb1564aecb31495dd4c65727" Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.160983 4884 scope.go:117] "RemoveContainer" containerID="e9822b1e0d3d2da8a4e8ca7a2775a0b81fad15877d18acf4f6666c01dbb7357f" Dec 02 03:09:47 crc kubenswrapper[4884]: E1202 03:09:47.174627 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9822b1e0d3d2da8a4e8ca7a2775a0b81fad15877d18acf4f6666c01dbb7357f\": container with ID starting with e9822b1e0d3d2da8a4e8ca7a2775a0b81fad15877d18acf4f6666c01dbb7357f not found: ID does not exist" containerID="e9822b1e0d3d2da8a4e8ca7a2775a0b81fad15877d18acf4f6666c01dbb7357f" Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.174669 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9822b1e0d3d2da8a4e8ca7a2775a0b81fad15877d18acf4f6666c01dbb7357f"} err="failed to get container status \"e9822b1e0d3d2da8a4e8ca7a2775a0b81fad15877d18acf4f6666c01dbb7357f\": rpc error: code = NotFound desc = could not find container \"e9822b1e0d3d2da8a4e8ca7a2775a0b81fad15877d18acf4f6666c01dbb7357f\": container with ID starting with e9822b1e0d3d2da8a4e8ca7a2775a0b81fad15877d18acf4f6666c01dbb7357f not found: ID does not exist" Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.174696 4884 scope.go:117] "RemoveContainer" containerID="2b962ef33a9791ef43af9d9bbff7e4b157a6c6f3de8a51852e259399f71d66d7" Dec 02 03:09:47 crc kubenswrapper[4884]: E1202 03:09:47.175855 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b962ef33a9791ef43af9d9bbff7e4b157a6c6f3de8a51852e259399f71d66d7\": container with ID starting with 2b962ef33a9791ef43af9d9bbff7e4b157a6c6f3de8a51852e259399f71d66d7 not found: ID does not exist" containerID="2b962ef33a9791ef43af9d9bbff7e4b157a6c6f3de8a51852e259399f71d66d7" Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.175985 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b962ef33a9791ef43af9d9bbff7e4b157a6c6f3de8a51852e259399f71d66d7"} err="failed to get container status \"2b962ef33a9791ef43af9d9bbff7e4b157a6c6f3de8a51852e259399f71d66d7\": rpc error: code = NotFound desc = could not find container \"2b962ef33a9791ef43af9d9bbff7e4b157a6c6f3de8a51852e259399f71d66d7\": container with ID starting with 2b962ef33a9791ef43af9d9bbff7e4b157a6c6f3de8a51852e259399f71d66d7 not found: ID does not exist" Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.175999 4884 scope.go:117] "RemoveContainer" containerID="718b60f5f8bf8d6cc02751a25b7054cf5ff59c44cb1564aecb31495dd4c65727" Dec 02 03:09:47 crc kubenswrapper[4884]: E1202 03:09:47.176290 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"718b60f5f8bf8d6cc02751a25b7054cf5ff59c44cb1564aecb31495dd4c65727\": container with ID starting with 718b60f5f8bf8d6cc02751a25b7054cf5ff59c44cb1564aecb31495dd4c65727 not found: ID does not exist" containerID="718b60f5f8bf8d6cc02751a25b7054cf5ff59c44cb1564aecb31495dd4c65727" Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.176311 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"718b60f5f8bf8d6cc02751a25b7054cf5ff59c44cb1564aecb31495dd4c65727"} err="failed to get container status \"718b60f5f8bf8d6cc02751a25b7054cf5ff59c44cb1564aecb31495dd4c65727\": rpc error: code = NotFound desc = could not find container \"718b60f5f8bf8d6cc02751a25b7054cf5ff59c44cb1564aecb31495dd4c65727\": container with ID starting with 718b60f5f8bf8d6cc02751a25b7054cf5ff59c44cb1564aecb31495dd4c65727 not found: ID does not exist" Dec 02 03:09:47 crc kubenswrapper[4884]: I1202 03:09:47.626209 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef66ab1f-432e-4274-a31b-68e8690c6f00" path="/var/lib/kubelet/pods/ef66ab1f-432e-4274-a31b-68e8690c6f00/volumes" Dec 02 03:10:04 crc kubenswrapper[4884]: I1202 03:10:04.631207 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zq7lm_e48fed54-40c5-40c5-a88a-71662222bb9e/control-plane-machine-set-operator/0.log" Dec 02 03:10:04 crc kubenswrapper[4884]: I1202 03:10:04.796936 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4htsc_2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4/machine-api-operator/0.log" Dec 02 03:10:04 crc kubenswrapper[4884]: I1202 03:10:04.813212 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-4htsc_2b7e2531-5e19-4cf0-9ff1-20e13b3db8b4/kube-rbac-proxy/0.log" Dec 02 03:10:19 crc kubenswrapper[4884]: I1202 03:10:19.294791 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-drxm6_ece97af9-2c98-424d-b314-f886b2c0cc7b/cert-manager-controller/0.log" Dec 02 03:10:19 crc kubenswrapper[4884]: I1202 03:10:19.462484 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-4kwx7_8b4e9534-399a-4836-8111-56f26c6a1b5d/cert-manager-cainjector/0.log" Dec 02 03:10:19 crc kubenswrapper[4884]: I1202 03:10:19.486703 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-6tp92_7708e2d7-d33c-4ffc-a56a-094109ba52ed/cert-manager-webhook/0.log" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.639628 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6dfrs"] Dec 02 03:10:31 crc kubenswrapper[4884]: E1202 03:10:31.642141 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef66ab1f-432e-4274-a31b-68e8690c6f00" containerName="extract-content" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.642505 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef66ab1f-432e-4274-a31b-68e8690c6f00" containerName="extract-content" Dec 02 03:10:31 crc kubenswrapper[4884]: E1202 03:10:31.642639 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef66ab1f-432e-4274-a31b-68e8690c6f00" containerName="extract-utilities" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.642777 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef66ab1f-432e-4274-a31b-68e8690c6f00" containerName="extract-utilities" Dec 02 03:10:31 crc kubenswrapper[4884]: E1202 03:10:31.642965 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef66ab1f-432e-4274-a31b-68e8690c6f00" containerName="registry-server" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.643046 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef66ab1f-432e-4274-a31b-68e8690c6f00" containerName="registry-server" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.643385 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef66ab1f-432e-4274-a31b-68e8690c6f00" containerName="registry-server" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.645301 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.658926 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6dfrs"] Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.807346 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-catalog-content\") pod \"certified-operators-6dfrs\" (UID: \"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3\") " pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.807867 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khps4\" (UniqueName: \"kubernetes.io/projected/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-kube-api-access-khps4\") pod \"certified-operators-6dfrs\" (UID: \"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3\") " pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.807966 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-utilities\") pod \"certified-operators-6dfrs\" (UID: \"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3\") " pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.910304 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-utilities\") pod \"certified-operators-6dfrs\" (UID: \"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3\") " pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.910492 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-catalog-content\") pod \"certified-operators-6dfrs\" (UID: \"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3\") " pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.910563 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khps4\" (UniqueName: \"kubernetes.io/projected/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-kube-api-access-khps4\") pod \"certified-operators-6dfrs\" (UID: \"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3\") " pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.911231 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-utilities\") pod \"certified-operators-6dfrs\" (UID: \"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3\") " pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.911378 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-catalog-content\") pod \"certified-operators-6dfrs\" (UID: \"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3\") " pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.946554 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khps4\" (UniqueName: \"kubernetes.io/projected/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-kube-api-access-khps4\") pod \"certified-operators-6dfrs\" (UID: \"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3\") " pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:31 crc kubenswrapper[4884]: I1202 03:10:31.986866 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:32 crc kubenswrapper[4884]: I1202 03:10:32.463257 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6dfrs"] Dec 02 03:10:32 crc kubenswrapper[4884]: I1202 03:10:32.578539 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6dfrs" event={"ID":"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3","Type":"ContainerStarted","Data":"94685686e163ca1a620603665075a2c02735026146d5c7136f5596a234819384"} Dec 02 03:10:33 crc kubenswrapper[4884]: I1202 03:10:33.589006 4884 generic.go:334] "Generic (PLEG): container finished" podID="ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3" containerID="c0946a8653e80f9e0e3f76461809fbdbcdc98677b317d2f959dce8703a5ce8ba" exitCode=0 Dec 02 03:10:33 crc kubenswrapper[4884]: I1202 03:10:33.589051 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6dfrs" event={"ID":"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3","Type":"ContainerDied","Data":"c0946a8653e80f9e0e3f76461809fbdbcdc98677b317d2f959dce8703a5ce8ba"} Dec 02 03:10:34 crc kubenswrapper[4884]: I1202 03:10:34.602541 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6dfrs" event={"ID":"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3","Type":"ContainerStarted","Data":"30c3160460a11c07c0793b4294722b98eaa5b97301af8844316810ae93de4106"} Dec 02 03:10:35 crc kubenswrapper[4884]: I1202 03:10:35.407658 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-p7chb"] Dec 02 03:10:35 crc kubenswrapper[4884]: I1202 03:10:35.410119 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:35 crc kubenswrapper[4884]: I1202 03:10:35.417311 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p7chb"] Dec 02 03:10:35 crc kubenswrapper[4884]: I1202 03:10:35.491212 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89a76d1a-057d-4659-8566-65a6d56dabb7-catalog-content\") pod \"redhat-operators-p7chb\" (UID: \"89a76d1a-057d-4659-8566-65a6d56dabb7\") " pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:35 crc kubenswrapper[4884]: I1202 03:10:35.491685 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89a76d1a-057d-4659-8566-65a6d56dabb7-utilities\") pod \"redhat-operators-p7chb\" (UID: \"89a76d1a-057d-4659-8566-65a6d56dabb7\") " pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:35 crc kubenswrapper[4884]: I1202 03:10:35.491729 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksqfp\" (UniqueName: \"kubernetes.io/projected/89a76d1a-057d-4659-8566-65a6d56dabb7-kube-api-access-ksqfp\") pod \"redhat-operators-p7chb\" (UID: \"89a76d1a-057d-4659-8566-65a6d56dabb7\") " pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:35 crc kubenswrapper[4884]: I1202 03:10:35.593967 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksqfp\" (UniqueName: \"kubernetes.io/projected/89a76d1a-057d-4659-8566-65a6d56dabb7-kube-api-access-ksqfp\") pod \"redhat-operators-p7chb\" (UID: \"89a76d1a-057d-4659-8566-65a6d56dabb7\") " pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:35 crc kubenswrapper[4884]: I1202 03:10:35.594391 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89a76d1a-057d-4659-8566-65a6d56dabb7-catalog-content\") pod \"redhat-operators-p7chb\" (UID: \"89a76d1a-057d-4659-8566-65a6d56dabb7\") " pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:35 crc kubenswrapper[4884]: I1202 03:10:35.594855 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89a76d1a-057d-4659-8566-65a6d56dabb7-utilities\") pod \"redhat-operators-p7chb\" (UID: \"89a76d1a-057d-4659-8566-65a6d56dabb7\") " pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:35 crc kubenswrapper[4884]: I1202 03:10:35.594940 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89a76d1a-057d-4659-8566-65a6d56dabb7-catalog-content\") pod \"redhat-operators-p7chb\" (UID: \"89a76d1a-057d-4659-8566-65a6d56dabb7\") " pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:35 crc kubenswrapper[4884]: I1202 03:10:35.595541 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89a76d1a-057d-4659-8566-65a6d56dabb7-utilities\") pod \"redhat-operators-p7chb\" (UID: \"89a76d1a-057d-4659-8566-65a6d56dabb7\") " pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:35 crc kubenswrapper[4884]: I1202 03:10:35.616400 4884 generic.go:334] "Generic (PLEG): container finished" podID="ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3" containerID="30c3160460a11c07c0793b4294722b98eaa5b97301af8844316810ae93de4106" exitCode=0 Dec 02 03:10:35 crc kubenswrapper[4884]: I1202 03:10:35.628939 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6dfrs" event={"ID":"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3","Type":"ContainerDied","Data":"30c3160460a11c07c0793b4294722b98eaa5b97301af8844316810ae93de4106"} Dec 02 03:10:35 crc kubenswrapper[4884]: I1202 03:10:35.750661 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksqfp\" (UniqueName: \"kubernetes.io/projected/89a76d1a-057d-4659-8566-65a6d56dabb7-kube-api-access-ksqfp\") pod \"redhat-operators-p7chb\" (UID: \"89a76d1a-057d-4659-8566-65a6d56dabb7\") " pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:35 crc kubenswrapper[4884]: I1202 03:10:35.761459 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:36 crc kubenswrapper[4884]: I1202 03:10:36.231655 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-p7chb"] Dec 02 03:10:36 crc kubenswrapper[4884]: W1202 03:10:36.236583 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89a76d1a_057d_4659_8566_65a6d56dabb7.slice/crio-b94d55ad3f85d9b6249044abcfb4f6b8e9c85fd8eaa6987ba09ff0129ff6ad7f WatchSource:0}: Error finding container b94d55ad3f85d9b6249044abcfb4f6b8e9c85fd8eaa6987ba09ff0129ff6ad7f: Status 404 returned error can't find the container with id b94d55ad3f85d9b6249044abcfb4f6b8e9c85fd8eaa6987ba09ff0129ff6ad7f Dec 02 03:10:36 crc kubenswrapper[4884]: I1202 03:10:36.643830 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6dfrs" event={"ID":"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3","Type":"ContainerStarted","Data":"8b9ea6ccf30d080ca743032ef7cbe6e058430767bc04b67373bde0e06431cb32"} Dec 02 03:10:36 crc kubenswrapper[4884]: I1202 03:10:36.648489 4884 generic.go:334] "Generic (PLEG): container finished" podID="89a76d1a-057d-4659-8566-65a6d56dabb7" containerID="17cb65508d6d57b9b21c846b2a14bd257d755c1c52e014c9eb45539cb517b083" exitCode=0 Dec 02 03:10:36 crc kubenswrapper[4884]: I1202 03:10:36.648528 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p7chb" event={"ID":"89a76d1a-057d-4659-8566-65a6d56dabb7","Type":"ContainerDied","Data":"17cb65508d6d57b9b21c846b2a14bd257d755c1c52e014c9eb45539cb517b083"} Dec 02 03:10:36 crc kubenswrapper[4884]: I1202 03:10:36.648551 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p7chb" event={"ID":"89a76d1a-057d-4659-8566-65a6d56dabb7","Type":"ContainerStarted","Data":"b94d55ad3f85d9b6249044abcfb4f6b8e9c85fd8eaa6987ba09ff0129ff6ad7f"} Dec 02 03:10:36 crc kubenswrapper[4884]: I1202 03:10:36.666722 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6dfrs" podStartSLOduration=3.199398648 podStartE2EDuration="5.666701318s" podCreationTimestamp="2025-12-02 03:10:31 +0000 UTC" firstStartedPulling="2025-12-02 03:10:33.591594705 +0000 UTC m=+5650.267431589" lastFinishedPulling="2025-12-02 03:10:36.058897375 +0000 UTC m=+5652.734734259" observedRunningTime="2025-12-02 03:10:36.662003723 +0000 UTC m=+5653.337840607" watchObservedRunningTime="2025-12-02 03:10:36.666701318 +0000 UTC m=+5653.342538202" Dec 02 03:10:37 crc kubenswrapper[4884]: I1202 03:10:37.130685 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-6skt5_c8a2aedd-caa8-4a87-9429-9443ef9c0fe4/nmstate-console-plugin/0.log" Dec 02 03:10:37 crc kubenswrapper[4884]: I1202 03:10:37.682431 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-wmss4_4a1eae32-edb9-42c6-9a22-62012c26bf23/kube-rbac-proxy/0.log" Dec 02 03:10:37 crc kubenswrapper[4884]: I1202 03:10:37.696290 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p7chb" event={"ID":"89a76d1a-057d-4659-8566-65a6d56dabb7","Type":"ContainerStarted","Data":"af49a92fdfb73443da10c254b5971b98c54d938ebbbe719b41aeaa4eca0cc0bc"} Dec 02 03:10:37 crc kubenswrapper[4884]: I1202 03:10:37.703244 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-wmss4_4a1eae32-edb9-42c6-9a22-62012c26bf23/nmstate-metrics/0.log" Dec 02 03:10:37 crc kubenswrapper[4884]: I1202 03:10:37.721048 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-d5d28_51729f65-2f0f-4a8a-bf7d-1b91f706aa82/nmstate-handler/0.log" Dec 02 03:10:37 crc kubenswrapper[4884]: I1202 03:10:37.883855 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-wm7q4_7f575e02-628b-4380-a9d3-126a6ababd10/nmstate-operator/0.log" Dec 02 03:10:38 crc kubenswrapper[4884]: I1202 03:10:38.063401 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-2mzbq_c783fcfc-44e0-4a67-bc02-f4d0f45f10a9/nmstate-webhook/0.log" Dec 02 03:10:40 crc kubenswrapper[4884]: I1202 03:10:40.732204 4884 generic.go:334] "Generic (PLEG): container finished" podID="89a76d1a-057d-4659-8566-65a6d56dabb7" containerID="af49a92fdfb73443da10c254b5971b98c54d938ebbbe719b41aeaa4eca0cc0bc" exitCode=0 Dec 02 03:10:40 crc kubenswrapper[4884]: I1202 03:10:40.732258 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p7chb" event={"ID":"89a76d1a-057d-4659-8566-65a6d56dabb7","Type":"ContainerDied","Data":"af49a92fdfb73443da10c254b5971b98c54d938ebbbe719b41aeaa4eca0cc0bc"} Dec 02 03:10:41 crc kubenswrapper[4884]: I1202 03:10:41.766404 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p7chb" event={"ID":"89a76d1a-057d-4659-8566-65a6d56dabb7","Type":"ContainerStarted","Data":"fd46fcff82c1871bce407cf6a7b9a4e396984e36c596281b0971a9f4d70e01ce"} Dec 02 03:10:41 crc kubenswrapper[4884]: I1202 03:10:41.789849 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-p7chb" podStartSLOduration=2.19973622 podStartE2EDuration="6.789835053s" podCreationTimestamp="2025-12-02 03:10:35 +0000 UTC" firstStartedPulling="2025-12-02 03:10:36.650081641 +0000 UTC m=+5653.325918525" lastFinishedPulling="2025-12-02 03:10:41.240180474 +0000 UTC m=+5657.916017358" observedRunningTime="2025-12-02 03:10:41.779650264 +0000 UTC m=+5658.455487158" watchObservedRunningTime="2025-12-02 03:10:41.789835053 +0000 UTC m=+5658.465671937" Dec 02 03:10:41 crc kubenswrapper[4884]: I1202 03:10:41.987805 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:41 crc kubenswrapper[4884]: I1202 03:10:41.988214 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:42 crc kubenswrapper[4884]: I1202 03:10:42.038831 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:42 crc kubenswrapper[4884]: I1202 03:10:42.828808 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:44 crc kubenswrapper[4884]: I1202 03:10:44.194493 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6dfrs"] Dec 02 03:10:45 crc kubenswrapper[4884]: I1202 03:10:45.761789 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:45 crc kubenswrapper[4884]: I1202 03:10:45.762128 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:45 crc kubenswrapper[4884]: I1202 03:10:45.808363 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6dfrs" podUID="ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3" containerName="registry-server" containerID="cri-o://8b9ea6ccf30d080ca743032ef7cbe6e058430767bc04b67373bde0e06431cb32" gracePeriod=2 Dec 02 03:10:46 crc kubenswrapper[4884]: I1202 03:10:46.292864 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:46 crc kubenswrapper[4884]: I1202 03:10:46.360424 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khps4\" (UniqueName: \"kubernetes.io/projected/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-kube-api-access-khps4\") pod \"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3\" (UID: \"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3\") " Dec 02 03:10:46 crc kubenswrapper[4884]: I1202 03:10:46.360541 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-catalog-content\") pod \"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3\" (UID: \"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3\") " Dec 02 03:10:46 crc kubenswrapper[4884]: I1202 03:10:46.360637 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-utilities\") pod \"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3\" (UID: \"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3\") " Dec 02 03:10:46 crc kubenswrapper[4884]: I1202 03:10:46.361379 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-utilities" (OuterVolumeSpecName: "utilities") pod "ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3" (UID: "ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 03:10:46 crc kubenswrapper[4884]: I1202 03:10:46.406446 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3" (UID: "ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 03:10:46 crc kubenswrapper[4884]: I1202 03:10:46.462443 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 03:10:46 crc kubenswrapper[4884]: I1202 03:10:46.462490 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 03:10:46 crc kubenswrapper[4884]: I1202 03:10:46.829249 4884 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-p7chb" podUID="89a76d1a-057d-4659-8566-65a6d56dabb7" containerName="registry-server" probeResult="failure" output=< Dec 02 03:10:46 crc kubenswrapper[4884]: timeout: failed to connect service ":50051" within 1s Dec 02 03:10:46 crc kubenswrapper[4884]: > Dec 02 03:10:46 crc kubenswrapper[4884]: I1202 03:10:46.830208 4884 generic.go:334] "Generic (PLEG): container finished" podID="ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3" containerID="8b9ea6ccf30d080ca743032ef7cbe6e058430767bc04b67373bde0e06431cb32" exitCode=0 Dec 02 03:10:46 crc kubenswrapper[4884]: I1202 03:10:46.830247 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6dfrs" event={"ID":"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3","Type":"ContainerDied","Data":"8b9ea6ccf30d080ca743032ef7cbe6e058430767bc04b67373bde0e06431cb32"} Dec 02 03:10:46 crc kubenswrapper[4884]: I1202 03:10:46.830273 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6dfrs" event={"ID":"ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3","Type":"ContainerDied","Data":"94685686e163ca1a620603665075a2c02735026146d5c7136f5596a234819384"} Dec 02 03:10:46 crc kubenswrapper[4884]: I1202 03:10:46.830289 4884 scope.go:117] "RemoveContainer" containerID="8b9ea6ccf30d080ca743032ef7cbe6e058430767bc04b67373bde0e06431cb32" Dec 02 03:10:46 crc kubenswrapper[4884]: I1202 03:10:46.830347 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6dfrs" Dec 02 03:10:46 crc kubenswrapper[4884]: I1202 03:10:46.868601 4884 scope.go:117] "RemoveContainer" containerID="30c3160460a11c07c0793b4294722b98eaa5b97301af8844316810ae93de4106" Dec 02 03:10:47 crc kubenswrapper[4884]: I1202 03:10:47.072702 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-kube-api-access-khps4" (OuterVolumeSpecName: "kube-api-access-khps4") pod "ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3" (UID: "ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3"). InnerVolumeSpecName "kube-api-access-khps4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 03:10:47 crc kubenswrapper[4884]: I1202 03:10:47.074635 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-khps4\" (UniqueName: \"kubernetes.io/projected/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3-kube-api-access-khps4\") on node \"crc\" DevicePath \"\"" Dec 02 03:10:47 crc kubenswrapper[4884]: I1202 03:10:47.104854 4884 scope.go:117] "RemoveContainer" containerID="c0946a8653e80f9e0e3f76461809fbdbcdc98677b317d2f959dce8703a5ce8ba" Dec 02 03:10:47 crc kubenswrapper[4884]: I1202 03:10:47.193088 4884 scope.go:117] "RemoveContainer" containerID="8b9ea6ccf30d080ca743032ef7cbe6e058430767bc04b67373bde0e06431cb32" Dec 02 03:10:47 crc kubenswrapper[4884]: E1202 03:10:47.193569 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b9ea6ccf30d080ca743032ef7cbe6e058430767bc04b67373bde0e06431cb32\": container with ID starting with 8b9ea6ccf30d080ca743032ef7cbe6e058430767bc04b67373bde0e06431cb32 not found: ID does not exist" containerID="8b9ea6ccf30d080ca743032ef7cbe6e058430767bc04b67373bde0e06431cb32" Dec 02 03:10:47 crc kubenswrapper[4884]: I1202 03:10:47.193602 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b9ea6ccf30d080ca743032ef7cbe6e058430767bc04b67373bde0e06431cb32"} err="failed to get container status \"8b9ea6ccf30d080ca743032ef7cbe6e058430767bc04b67373bde0e06431cb32\": rpc error: code = NotFound desc = could not find container \"8b9ea6ccf30d080ca743032ef7cbe6e058430767bc04b67373bde0e06431cb32\": container with ID starting with 8b9ea6ccf30d080ca743032ef7cbe6e058430767bc04b67373bde0e06431cb32 not found: ID does not exist" Dec 02 03:10:47 crc kubenswrapper[4884]: I1202 03:10:47.193624 4884 scope.go:117] "RemoveContainer" containerID="30c3160460a11c07c0793b4294722b98eaa5b97301af8844316810ae93de4106" Dec 02 03:10:47 crc kubenswrapper[4884]: E1202 03:10:47.193984 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30c3160460a11c07c0793b4294722b98eaa5b97301af8844316810ae93de4106\": container with ID starting with 30c3160460a11c07c0793b4294722b98eaa5b97301af8844316810ae93de4106 not found: ID does not exist" containerID="30c3160460a11c07c0793b4294722b98eaa5b97301af8844316810ae93de4106" Dec 02 03:10:47 crc kubenswrapper[4884]: I1202 03:10:47.194036 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30c3160460a11c07c0793b4294722b98eaa5b97301af8844316810ae93de4106"} err="failed to get container status \"30c3160460a11c07c0793b4294722b98eaa5b97301af8844316810ae93de4106\": rpc error: code = NotFound desc = could not find container \"30c3160460a11c07c0793b4294722b98eaa5b97301af8844316810ae93de4106\": container with ID starting with 30c3160460a11c07c0793b4294722b98eaa5b97301af8844316810ae93de4106 not found: ID does not exist" Dec 02 03:10:47 crc kubenswrapper[4884]: I1202 03:10:47.194069 4884 scope.go:117] "RemoveContainer" containerID="c0946a8653e80f9e0e3f76461809fbdbcdc98677b317d2f959dce8703a5ce8ba" Dec 02 03:10:47 crc kubenswrapper[4884]: E1202 03:10:47.194352 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0946a8653e80f9e0e3f76461809fbdbcdc98677b317d2f959dce8703a5ce8ba\": container with ID starting with c0946a8653e80f9e0e3f76461809fbdbcdc98677b317d2f959dce8703a5ce8ba not found: ID does not exist" containerID="c0946a8653e80f9e0e3f76461809fbdbcdc98677b317d2f959dce8703a5ce8ba" Dec 02 03:10:47 crc kubenswrapper[4884]: I1202 03:10:47.194375 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0946a8653e80f9e0e3f76461809fbdbcdc98677b317d2f959dce8703a5ce8ba"} err="failed to get container status \"c0946a8653e80f9e0e3f76461809fbdbcdc98677b317d2f959dce8703a5ce8ba\": rpc error: code = NotFound desc = could not find container \"c0946a8653e80f9e0e3f76461809fbdbcdc98677b317d2f959dce8703a5ce8ba\": container with ID starting with c0946a8653e80f9e0e3f76461809fbdbcdc98677b317d2f959dce8703a5ce8ba not found: ID does not exist" Dec 02 03:10:47 crc kubenswrapper[4884]: I1202 03:10:47.241516 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6dfrs"] Dec 02 03:10:47 crc kubenswrapper[4884]: I1202 03:10:47.252935 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6dfrs"] Dec 02 03:10:47 crc kubenswrapper[4884]: I1202 03:10:47.628900 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3" path="/var/lib/kubelet/pods/ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3/volumes" Dec 02 03:10:55 crc kubenswrapper[4884]: I1202 03:10:55.818216 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:55 crc kubenswrapper[4884]: I1202 03:10:55.865802 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:56 crc kubenswrapper[4884]: I1202 03:10:56.057326 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p7chb"] Dec 02 03:10:56 crc kubenswrapper[4884]: I1202 03:10:56.456122 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-8wv2n_e365201e-b871-4206-8dcd-9b87a9452f83/kube-rbac-proxy/0.log" Dec 02 03:10:56 crc kubenswrapper[4884]: I1202 03:10:56.539945 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-8wv2n_e365201e-b871-4206-8dcd-9b87a9452f83/controller/0.log" Dec 02 03:10:56 crc kubenswrapper[4884]: I1202 03:10:56.934525 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-p7chb" podUID="89a76d1a-057d-4659-8566-65a6d56dabb7" containerName="registry-server" containerID="cri-o://fd46fcff82c1871bce407cf6a7b9a4e396984e36c596281b0971a9f4d70e01ce" gracePeriod=2 Dec 02 03:10:57 crc kubenswrapper[4884]: I1202 03:10:57.115982 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/cp-frr-files/0.log" Dec 02 03:10:57 crc kubenswrapper[4884]: I1202 03:10:57.286282 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/cp-frr-files/0.log" Dec 02 03:10:57 crc kubenswrapper[4884]: I1202 03:10:57.332797 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/cp-metrics/0.log" Dec 02 03:10:57 crc kubenswrapper[4884]: I1202 03:10:57.346383 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/cp-reloader/0.log" Dec 02 03:10:57 crc kubenswrapper[4884]: I1202 03:10:57.392050 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/cp-reloader/0.log" Dec 02 03:10:57 crc kubenswrapper[4884]: I1202 03:10:57.566642 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/cp-frr-files/0.log" Dec 02 03:10:57 crc kubenswrapper[4884]: I1202 03:10:57.605409 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/cp-metrics/0.log" Dec 02 03:10:57 crc kubenswrapper[4884]: I1202 03:10:57.608060 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/cp-reloader/0.log" Dec 02 03:10:57 crc kubenswrapper[4884]: I1202 03:10:57.653473 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/cp-metrics/0.log" Dec 02 03:10:57 crc kubenswrapper[4884]: I1202 03:10:57.850800 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/cp-reloader/0.log" Dec 02 03:10:57 crc kubenswrapper[4884]: I1202 03:10:57.925440 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/cp-frr-files/0.log" Dec 02 03:10:57 crc kubenswrapper[4884]: I1202 03:10:57.931450 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/cp-metrics/0.log" Dec 02 03:10:57 crc kubenswrapper[4884]: I1202 03:10:57.946460 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/controller/0.log" Dec 02 03:10:57 crc kubenswrapper[4884]: I1202 03:10:57.954506 4884 generic.go:334] "Generic (PLEG): container finished" podID="89a76d1a-057d-4659-8566-65a6d56dabb7" containerID="fd46fcff82c1871bce407cf6a7b9a4e396984e36c596281b0971a9f4d70e01ce" exitCode=0 Dec 02 03:10:57 crc kubenswrapper[4884]: I1202 03:10:57.954547 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p7chb" event={"ID":"89a76d1a-057d-4659-8566-65a6d56dabb7","Type":"ContainerDied","Data":"fd46fcff82c1871bce407cf6a7b9a4e396984e36c596281b0971a9f4d70e01ce"} Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.382235 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.497376 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89a76d1a-057d-4659-8566-65a6d56dabb7-utilities\") pod \"89a76d1a-057d-4659-8566-65a6d56dabb7\" (UID: \"89a76d1a-057d-4659-8566-65a6d56dabb7\") " Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.497541 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89a76d1a-057d-4659-8566-65a6d56dabb7-catalog-content\") pod \"89a76d1a-057d-4659-8566-65a6d56dabb7\" (UID: \"89a76d1a-057d-4659-8566-65a6d56dabb7\") " Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.497596 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksqfp\" (UniqueName: \"kubernetes.io/projected/89a76d1a-057d-4659-8566-65a6d56dabb7-kube-api-access-ksqfp\") pod \"89a76d1a-057d-4659-8566-65a6d56dabb7\" (UID: \"89a76d1a-057d-4659-8566-65a6d56dabb7\") " Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.497966 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89a76d1a-057d-4659-8566-65a6d56dabb7-utilities" (OuterVolumeSpecName: "utilities") pod "89a76d1a-057d-4659-8566-65a6d56dabb7" (UID: "89a76d1a-057d-4659-8566-65a6d56dabb7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.509831 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89a76d1a-057d-4659-8566-65a6d56dabb7-kube-api-access-ksqfp" (OuterVolumeSpecName: "kube-api-access-ksqfp") pod "89a76d1a-057d-4659-8566-65a6d56dabb7" (UID: "89a76d1a-057d-4659-8566-65a6d56dabb7"). InnerVolumeSpecName "kube-api-access-ksqfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.527437 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/kube-rbac-proxy-frr/0.log" Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.536346 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/kube-rbac-proxy/0.log" Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.579595 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/frr-metrics/0.log" Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.599338 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksqfp\" (UniqueName: \"kubernetes.io/projected/89a76d1a-057d-4659-8566-65a6d56dabb7-kube-api-access-ksqfp\") on node \"crc\" DevicePath \"\"" Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.599370 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89a76d1a-057d-4659-8566-65a6d56dabb7-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.614339 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89a76d1a-057d-4659-8566-65a6d56dabb7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89a76d1a-057d-4659-8566-65a6d56dabb7" (UID: "89a76d1a-057d-4659-8566-65a6d56dabb7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.700570 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89a76d1a-057d-4659-8566-65a6d56dabb7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.767974 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-5j9vm_7ef4b609-a02d-4050-a5cc-768308bf2dcb/frr-k8s-webhook-server/0.log" Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.793958 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/reloader/0.log" Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.975327 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-p7chb" event={"ID":"89a76d1a-057d-4659-8566-65a6d56dabb7","Type":"ContainerDied","Data":"b94d55ad3f85d9b6249044abcfb4f6b8e9c85fd8eaa6987ba09ff0129ff6ad7f"} Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.975372 4884 scope.go:117] "RemoveContainer" containerID="fd46fcff82c1871bce407cf6a7b9a4e396984e36c596281b0971a9f4d70e01ce" Dec 02 03:10:58 crc kubenswrapper[4884]: I1202 03:10:58.975502 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-p7chb" Dec 02 03:10:59 crc kubenswrapper[4884]: I1202 03:10:59.008916 4884 scope.go:117] "RemoveContainer" containerID="af49a92fdfb73443da10c254b5971b98c54d938ebbbe719b41aeaa4eca0cc0bc" Dec 02 03:10:59 crc kubenswrapper[4884]: I1202 03:10:59.067046 4884 scope.go:117] "RemoveContainer" containerID="17cb65508d6d57b9b21c846b2a14bd257d755c1c52e014c9eb45539cb517b083" Dec 02 03:10:59 crc kubenswrapper[4884]: I1202 03:10:59.091794 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-p7chb"] Dec 02 03:10:59 crc kubenswrapper[4884]: I1202 03:10:59.127707 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-p7chb"] Dec 02 03:10:59 crc kubenswrapper[4884]: I1202 03:10:59.152971 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-864859b84c-8pkv6_ebe290f2-7fe0-43c7-9485-a7d1789a4460/manager/0.log" Dec 02 03:10:59 crc kubenswrapper[4884]: E1202 03:10:59.285270 4884 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89a76d1a_057d_4659_8566_65a6d56dabb7.slice\": RecentStats: unable to find data in memory cache]" Dec 02 03:10:59 crc kubenswrapper[4884]: I1202 03:10:59.373789 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6c4c8b8bbb-plfsx_cf133a27-1ee3-4426-b67a-dd58ce471d16/webhook-server/0.log" Dec 02 03:10:59 crc kubenswrapper[4884]: I1202 03:10:59.417251 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bxtbh_f3fc51b3-99fe-4f7f-be90-68acd111b7ca/kube-rbac-proxy/0.log" Dec 02 03:10:59 crc kubenswrapper[4884]: I1202 03:10:59.624928 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89a76d1a-057d-4659-8566-65a6d56dabb7" path="/var/lib/kubelet/pods/89a76d1a-057d-4659-8566-65a6d56dabb7/volumes" Dec 02 03:10:59 crc kubenswrapper[4884]: I1202 03:10:59.953287 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-bxtbh_f3fc51b3-99fe-4f7f-be90-68acd111b7ca/speaker/0.log" Dec 02 03:11:00 crc kubenswrapper[4884]: I1202 03:11:00.103008 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-j7xs4_d0e2e152-8ae0-495d-835c-5dca5ef66b6c/frr/0.log" Dec 02 03:11:15 crc kubenswrapper[4884]: I1202 03:11:15.835459 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872_b885e1cf-71b5-4b0f-8a59-d213d14db08a/util/0.log" Dec 02 03:11:16 crc kubenswrapper[4884]: I1202 03:11:16.069523 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872_b885e1cf-71b5-4b0f-8a59-d213d14db08a/pull/0.log" Dec 02 03:11:16 crc kubenswrapper[4884]: I1202 03:11:16.077128 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872_b885e1cf-71b5-4b0f-8a59-d213d14db08a/pull/0.log" Dec 02 03:11:16 crc kubenswrapper[4884]: I1202 03:11:16.094969 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872_b885e1cf-71b5-4b0f-8a59-d213d14db08a/util/0.log" Dec 02 03:11:16 crc kubenswrapper[4884]: I1202 03:11:16.205622 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872_b885e1cf-71b5-4b0f-8a59-d213d14db08a/util/0.log" Dec 02 03:11:16 crc kubenswrapper[4884]: I1202 03:11:16.245869 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872_b885e1cf-71b5-4b0f-8a59-d213d14db08a/extract/0.log" Dec 02 03:11:16 crc kubenswrapper[4884]: I1202 03:11:16.249304 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fdt872_b885e1cf-71b5-4b0f-8a59-d213d14db08a/pull/0.log" Dec 02 03:11:16 crc kubenswrapper[4884]: I1202 03:11:16.382544 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_f4a3aaad-0061-494c-a330-c7484a1a5108/util/0.log" Dec 02 03:11:16 crc kubenswrapper[4884]: I1202 03:11:16.549944 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_f4a3aaad-0061-494c-a330-c7484a1a5108/util/0.log" Dec 02 03:11:16 crc kubenswrapper[4884]: I1202 03:11:16.585693 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_f4a3aaad-0061-494c-a330-c7484a1a5108/pull/0.log" Dec 02 03:11:16 crc kubenswrapper[4884]: I1202 03:11:16.603562 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_f4a3aaad-0061-494c-a330-c7484a1a5108/pull/0.log" Dec 02 03:11:16 crc kubenswrapper[4884]: I1202 03:11:16.829225 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_f4a3aaad-0061-494c-a330-c7484a1a5108/extract/0.log" Dec 02 03:11:16 crc kubenswrapper[4884]: I1202 03:11:16.832704 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_f4a3aaad-0061-494c-a330-c7484a1a5108/util/0.log" Dec 02 03:11:16 crc kubenswrapper[4884]: I1202 03:11:16.872578 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tp9xm_f4a3aaad-0061-494c-a330-c7484a1a5108/pull/0.log" Dec 02 03:11:17 crc kubenswrapper[4884]: I1202 03:11:17.006122 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t_2019f363-8aa0-4167-8fdc-ebf932fd07ed/util/0.log" Dec 02 03:11:17 crc kubenswrapper[4884]: I1202 03:11:17.314838 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t_2019f363-8aa0-4167-8fdc-ebf932fd07ed/util/0.log" Dec 02 03:11:17 crc kubenswrapper[4884]: I1202 03:11:17.331511 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t_2019f363-8aa0-4167-8fdc-ebf932fd07ed/pull/0.log" Dec 02 03:11:17 crc kubenswrapper[4884]: I1202 03:11:17.332884 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t_2019f363-8aa0-4167-8fdc-ebf932fd07ed/pull/0.log" Dec 02 03:11:17 crc kubenswrapper[4884]: I1202 03:11:17.602244 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t_2019f363-8aa0-4167-8fdc-ebf932fd07ed/pull/0.log" Dec 02 03:11:17 crc kubenswrapper[4884]: I1202 03:11:17.648785 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t_2019f363-8aa0-4167-8fdc-ebf932fd07ed/util/0.log" Dec 02 03:11:17 crc kubenswrapper[4884]: I1202 03:11:17.654059 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83bd69t_2019f363-8aa0-4167-8fdc-ebf932fd07ed/extract/0.log" Dec 02 03:11:17 crc kubenswrapper[4884]: I1202 03:11:17.798661 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jspph_a80d84c1-0c7c-44ff-b7f2-fe4704103241/extract-utilities/0.log" Dec 02 03:11:17 crc kubenswrapper[4884]: I1202 03:11:17.989756 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jspph_a80d84c1-0c7c-44ff-b7f2-fe4704103241/extract-content/0.log" Dec 02 03:11:17 crc kubenswrapper[4884]: I1202 03:11:17.998308 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jspph_a80d84c1-0c7c-44ff-b7f2-fe4704103241/extract-utilities/0.log" Dec 02 03:11:18 crc kubenswrapper[4884]: I1202 03:11:18.056714 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jspph_a80d84c1-0c7c-44ff-b7f2-fe4704103241/extract-content/0.log" Dec 02 03:11:18 crc kubenswrapper[4884]: I1202 03:11:18.156247 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jspph_a80d84c1-0c7c-44ff-b7f2-fe4704103241/extract-utilities/0.log" Dec 02 03:11:18 crc kubenswrapper[4884]: I1202 03:11:18.226920 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jspph_a80d84c1-0c7c-44ff-b7f2-fe4704103241/extract-content/0.log" Dec 02 03:11:18 crc kubenswrapper[4884]: I1202 03:11:18.374624 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p4sd7_154cc736-2f3e-4254-835f-c32fe8557341/extract-utilities/0.log" Dec 02 03:11:18 crc kubenswrapper[4884]: I1202 03:11:18.619423 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p4sd7_154cc736-2f3e-4254-835f-c32fe8557341/extract-content/0.log" Dec 02 03:11:18 crc kubenswrapper[4884]: I1202 03:11:18.642659 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p4sd7_154cc736-2f3e-4254-835f-c32fe8557341/extract-content/0.log" Dec 02 03:11:18 crc kubenswrapper[4884]: I1202 03:11:18.683165 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p4sd7_154cc736-2f3e-4254-835f-c32fe8557341/extract-utilities/0.log" Dec 02 03:11:18 crc kubenswrapper[4884]: I1202 03:11:18.900598 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jspph_a80d84c1-0c7c-44ff-b7f2-fe4704103241/registry-server/0.log" Dec 02 03:11:18 crc kubenswrapper[4884]: I1202 03:11:18.959379 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p4sd7_154cc736-2f3e-4254-835f-c32fe8557341/extract-content/0.log" Dec 02 03:11:18 crc kubenswrapper[4884]: I1202 03:11:18.981680 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p4sd7_154cc736-2f3e-4254-835f-c32fe8557341/extract-utilities/0.log" Dec 02 03:11:19 crc kubenswrapper[4884]: I1202 03:11:19.186993 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-gcw7d_ce86b3f0-07c2-4c81-91e0-d0b51cc0ece1/marketplace-operator/0.log" Dec 02 03:11:19 crc kubenswrapper[4884]: I1202 03:11:19.263468 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dddxc_cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85/extract-utilities/0.log" Dec 02 03:11:19 crc kubenswrapper[4884]: I1202 03:11:19.576934 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dddxc_cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85/extract-content/0.log" Dec 02 03:11:19 crc kubenswrapper[4884]: I1202 03:11:19.684411 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p4sd7_154cc736-2f3e-4254-835f-c32fe8557341/registry-server/0.log" Dec 02 03:11:19 crc kubenswrapper[4884]: I1202 03:11:19.800964 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dddxc_cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85/extract-content/0.log" Dec 02 03:11:19 crc kubenswrapper[4884]: I1202 03:11:19.812008 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dddxc_cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85/extract-utilities/0.log" Dec 02 03:11:20 crc kubenswrapper[4884]: I1202 03:11:20.064806 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dddxc_cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85/extract-content/0.log" Dec 02 03:11:20 crc kubenswrapper[4884]: I1202 03:11:20.067302 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bmrjj_81df2b19-2e90-4a01-b333-b4b072229d90/extract-utilities/0.log" Dec 02 03:11:20 crc kubenswrapper[4884]: I1202 03:11:20.080149 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dddxc_cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85/extract-utilities/0.log" Dec 02 03:11:20 crc kubenswrapper[4884]: I1202 03:11:20.320351 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-dddxc_cd3c5eb9-e3ac-431f-88ee-916b5d9a7b85/registry-server/0.log" Dec 02 03:11:20 crc kubenswrapper[4884]: I1202 03:11:20.346455 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bmrjj_81df2b19-2e90-4a01-b333-b4b072229d90/extract-utilities/0.log" Dec 02 03:11:20 crc kubenswrapper[4884]: I1202 03:11:20.346621 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bmrjj_81df2b19-2e90-4a01-b333-b4b072229d90/extract-content/0.log" Dec 02 03:11:20 crc kubenswrapper[4884]: I1202 03:11:20.367165 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bmrjj_81df2b19-2e90-4a01-b333-b4b072229d90/extract-content/0.log" Dec 02 03:11:21 crc kubenswrapper[4884]: I1202 03:11:21.149130 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bmrjj_81df2b19-2e90-4a01-b333-b4b072229d90/extract-utilities/0.log" Dec 02 03:11:21 crc kubenswrapper[4884]: I1202 03:11:21.221958 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bmrjj_81df2b19-2e90-4a01-b333-b4b072229d90/extract-content/0.log" Dec 02 03:11:21 crc kubenswrapper[4884]: I1202 03:11:21.793996 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bmrjj_81df2b19-2e90-4a01-b333-b4b072229d90/registry-server/0.log" Dec 02 03:11:37 crc kubenswrapper[4884]: I1202 03:11:37.082755 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-fbzpt_d1dc28c3-894b-4a33-98fa-bc56f4ba69e7/prometheus-operator/0.log" Dec 02 03:11:37 crc kubenswrapper[4884]: I1202 03:11:37.275820 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-86546477d5-9m7lr_6175052e-784f-4b4a-8fc0-7f26d6750a37/prometheus-operator-admission-webhook/0.log" Dec 02 03:11:37 crc kubenswrapper[4884]: I1202 03:11:37.325448 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-86546477d5-lclj9_d443ca5c-ecf4-49ba-97fc-faccc445af05/prometheus-operator-admission-webhook/0.log" Dec 02 03:11:37 crc kubenswrapper[4884]: I1202 03:11:37.694951 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-t8m6m_5624efce-3f02-447f-907f-4041a8a86629/operator/0.log" Dec 02 03:11:37 crc kubenswrapper[4884]: I1202 03:11:37.783677 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-h2v4b_a5fac10c-4a51-4406-a692-9020732cc0a4/perses-operator/0.log" Dec 02 03:11:46 crc kubenswrapper[4884]: I1202 03:11:46.971479 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 03:11:46 crc kubenswrapper[4884]: I1202 03:11:46.971986 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 03:12:16 crc kubenswrapper[4884]: I1202 03:12:16.972245 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 03:12:16 crc kubenswrapper[4884]: I1202 03:12:16.973010 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 03:12:46 crc kubenswrapper[4884]: I1202 03:12:46.972018 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 03:12:46 crc kubenswrapper[4884]: I1202 03:12:46.973184 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 03:12:46 crc kubenswrapper[4884]: I1202 03:12:46.973307 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 03:12:46 crc kubenswrapper[4884]: I1202 03:12:46.974837 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"783609d4dac9b374973e24fed69e04030080c197a1b2e0ea5d35a59a53de30f9"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 03:12:46 crc kubenswrapper[4884]: I1202 03:12:46.975001 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://783609d4dac9b374973e24fed69e04030080c197a1b2e0ea5d35a59a53de30f9" gracePeriod=600 Dec 02 03:12:47 crc kubenswrapper[4884]: I1202 03:12:47.194046 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="783609d4dac9b374973e24fed69e04030080c197a1b2e0ea5d35a59a53de30f9" exitCode=0 Dec 02 03:12:47 crc kubenswrapper[4884]: I1202 03:12:47.194160 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"783609d4dac9b374973e24fed69e04030080c197a1b2e0ea5d35a59a53de30f9"} Dec 02 03:12:47 crc kubenswrapper[4884]: I1202 03:12:47.194451 4884 scope.go:117] "RemoveContainer" containerID="bc6369a225a8d0a6d62326fcf9892d8031a569d63d4d0a6abf3d9edc1b88166a" Dec 02 03:12:48 crc kubenswrapper[4884]: I1202 03:12:48.208826 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerStarted","Data":"056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc"} Dec 02 03:13:27 crc kubenswrapper[4884]: I1202 03:13:27.790484 4884 generic.go:334] "Generic (PLEG): container finished" podID="284a08f7-5dcd-48f1-823e-2c4b4e8934da" containerID="1c40f6b01024689662e8040fd9156f4deff736d760bea7af7162c3b432ba8266" exitCode=0 Dec 02 03:13:27 crc kubenswrapper[4884]: I1202 03:13:27.790621 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h4s98/must-gather-vwhm4" event={"ID":"284a08f7-5dcd-48f1-823e-2c4b4e8934da","Type":"ContainerDied","Data":"1c40f6b01024689662e8040fd9156f4deff736d760bea7af7162c3b432ba8266"} Dec 02 03:13:27 crc kubenswrapper[4884]: I1202 03:13:27.792085 4884 scope.go:117] "RemoveContainer" containerID="1c40f6b01024689662e8040fd9156f4deff736d760bea7af7162c3b432ba8266" Dec 02 03:13:28 crc kubenswrapper[4884]: I1202 03:13:28.738075 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h4s98_must-gather-vwhm4_284a08f7-5dcd-48f1-823e-2c4b4e8934da/gather/0.log" Dec 02 03:13:37 crc kubenswrapper[4884]: I1202 03:13:37.713149 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-h4s98/must-gather-vwhm4"] Dec 02 03:13:37 crc kubenswrapper[4884]: I1202 03:13:37.715234 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-h4s98/must-gather-vwhm4" podUID="284a08f7-5dcd-48f1-823e-2c4b4e8934da" containerName="copy" containerID="cri-o://17508bad1f79c4bc670ee94895e9a60d9964781ab58bdd7355aa072eb25d5abd" gracePeriod=2 Dec 02 03:13:37 crc kubenswrapper[4884]: I1202 03:13:37.727502 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-h4s98/must-gather-vwhm4"] Dec 02 03:13:37 crc kubenswrapper[4884]: I1202 03:13:37.929450 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h4s98_must-gather-vwhm4_284a08f7-5dcd-48f1-823e-2c4b4e8934da/copy/0.log" Dec 02 03:13:37 crc kubenswrapper[4884]: I1202 03:13:37.930434 4884 generic.go:334] "Generic (PLEG): container finished" podID="284a08f7-5dcd-48f1-823e-2c4b4e8934da" containerID="17508bad1f79c4bc670ee94895e9a60d9964781ab58bdd7355aa072eb25d5abd" exitCode=143 Dec 02 03:13:38 crc kubenswrapper[4884]: I1202 03:13:38.205613 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h4s98_must-gather-vwhm4_284a08f7-5dcd-48f1-823e-2c4b4e8934da/copy/0.log" Dec 02 03:13:38 crc kubenswrapper[4884]: I1202 03:13:38.206036 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/must-gather-vwhm4" Dec 02 03:13:38 crc kubenswrapper[4884]: I1202 03:13:38.383320 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/284a08f7-5dcd-48f1-823e-2c4b4e8934da-must-gather-output\") pod \"284a08f7-5dcd-48f1-823e-2c4b4e8934da\" (UID: \"284a08f7-5dcd-48f1-823e-2c4b4e8934da\") " Dec 02 03:13:38 crc kubenswrapper[4884]: I1202 03:13:38.383797 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zdct\" (UniqueName: \"kubernetes.io/projected/284a08f7-5dcd-48f1-823e-2c4b4e8934da-kube-api-access-4zdct\") pod \"284a08f7-5dcd-48f1-823e-2c4b4e8934da\" (UID: \"284a08f7-5dcd-48f1-823e-2c4b4e8934da\") " Dec 02 03:13:38 crc kubenswrapper[4884]: I1202 03:13:38.396268 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/284a08f7-5dcd-48f1-823e-2c4b4e8934da-kube-api-access-4zdct" (OuterVolumeSpecName: "kube-api-access-4zdct") pod "284a08f7-5dcd-48f1-823e-2c4b4e8934da" (UID: "284a08f7-5dcd-48f1-823e-2c4b4e8934da"). InnerVolumeSpecName "kube-api-access-4zdct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 03:13:38 crc kubenswrapper[4884]: I1202 03:13:38.487708 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zdct\" (UniqueName: \"kubernetes.io/projected/284a08f7-5dcd-48f1-823e-2c4b4e8934da-kube-api-access-4zdct\") on node \"crc\" DevicePath \"\"" Dec 02 03:13:38 crc kubenswrapper[4884]: I1202 03:13:38.568842 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/284a08f7-5dcd-48f1-823e-2c4b4e8934da-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "284a08f7-5dcd-48f1-823e-2c4b4e8934da" (UID: "284a08f7-5dcd-48f1-823e-2c4b4e8934da"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 03:13:38 crc kubenswrapper[4884]: I1202 03:13:38.588687 4884 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/284a08f7-5dcd-48f1-823e-2c4b4e8934da-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 02 03:13:38 crc kubenswrapper[4884]: I1202 03:13:38.948279 4884 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h4s98_must-gather-vwhm4_284a08f7-5dcd-48f1-823e-2c4b4e8934da/copy/0.log" Dec 02 03:13:38 crc kubenswrapper[4884]: I1202 03:13:38.949190 4884 scope.go:117] "RemoveContainer" containerID="17508bad1f79c4bc670ee94895e9a60d9964781ab58bdd7355aa072eb25d5abd" Dec 02 03:13:38 crc kubenswrapper[4884]: I1202 03:13:38.949379 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h4s98/must-gather-vwhm4" Dec 02 03:13:38 crc kubenswrapper[4884]: I1202 03:13:38.995177 4884 scope.go:117] "RemoveContainer" containerID="1c40f6b01024689662e8040fd9156f4deff736d760bea7af7162c3b432ba8266" Dec 02 03:13:39 crc kubenswrapper[4884]: I1202 03:13:39.628661 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="284a08f7-5dcd-48f1-823e-2c4b4e8934da" path="/var/lib/kubelet/pods/284a08f7-5dcd-48f1-823e-2c4b4e8934da/volumes" Dec 02 03:13:48 crc kubenswrapper[4884]: I1202 03:13:48.142621 4884 scope.go:117] "RemoveContainer" containerID="1fab3ed3787bf5044ebb631ddad31dbd559aac510caa4fd1512b6d81ea9d0abe" Dec 02 03:14:28 crc kubenswrapper[4884]: I1202 03:14:28.935257 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t88cv"] Dec 02 03:14:28 crc kubenswrapper[4884]: E1202 03:14:28.936901 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3" containerName="extract-content" Dec 02 03:14:28 crc kubenswrapper[4884]: I1202 03:14:28.936934 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3" containerName="extract-content" Dec 02 03:14:28 crc kubenswrapper[4884]: E1202 03:14:28.936994 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3" containerName="extract-utilities" Dec 02 03:14:28 crc kubenswrapper[4884]: I1202 03:14:28.937013 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3" containerName="extract-utilities" Dec 02 03:14:28 crc kubenswrapper[4884]: E1202 03:14:28.937048 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3" containerName="registry-server" Dec 02 03:14:28 crc kubenswrapper[4884]: I1202 03:14:28.937067 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3" containerName="registry-server" Dec 02 03:14:28 crc kubenswrapper[4884]: E1202 03:14:28.937092 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89a76d1a-057d-4659-8566-65a6d56dabb7" containerName="extract-content" Dec 02 03:14:28 crc kubenswrapper[4884]: I1202 03:14:28.937108 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a76d1a-057d-4659-8566-65a6d56dabb7" containerName="extract-content" Dec 02 03:14:28 crc kubenswrapper[4884]: E1202 03:14:28.937131 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89a76d1a-057d-4659-8566-65a6d56dabb7" containerName="registry-server" Dec 02 03:14:28 crc kubenswrapper[4884]: I1202 03:14:28.937147 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a76d1a-057d-4659-8566-65a6d56dabb7" containerName="registry-server" Dec 02 03:14:28 crc kubenswrapper[4884]: E1202 03:14:28.937186 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="284a08f7-5dcd-48f1-823e-2c4b4e8934da" containerName="copy" Dec 02 03:14:28 crc kubenswrapper[4884]: I1202 03:14:28.937204 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="284a08f7-5dcd-48f1-823e-2c4b4e8934da" containerName="copy" Dec 02 03:14:28 crc kubenswrapper[4884]: E1202 03:14:28.937221 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="284a08f7-5dcd-48f1-823e-2c4b4e8934da" containerName="gather" Dec 02 03:14:28 crc kubenswrapper[4884]: I1202 03:14:28.937236 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="284a08f7-5dcd-48f1-823e-2c4b4e8934da" containerName="gather" Dec 02 03:14:28 crc kubenswrapper[4884]: E1202 03:14:28.937273 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89a76d1a-057d-4659-8566-65a6d56dabb7" containerName="extract-utilities" Dec 02 03:14:28 crc kubenswrapper[4884]: I1202 03:14:28.937290 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="89a76d1a-057d-4659-8566-65a6d56dabb7" containerName="extract-utilities" Dec 02 03:14:28 crc kubenswrapper[4884]: I1202 03:14:28.937737 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba5d6ebb-27be-48d3-b3ca-7b13ff3692e3" containerName="registry-server" Dec 02 03:14:28 crc kubenswrapper[4884]: I1202 03:14:28.937820 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="284a08f7-5dcd-48f1-823e-2c4b4e8934da" containerName="copy" Dec 02 03:14:28 crc kubenswrapper[4884]: I1202 03:14:28.937919 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="89a76d1a-057d-4659-8566-65a6d56dabb7" containerName="registry-server" Dec 02 03:14:28 crc kubenswrapper[4884]: I1202 03:14:28.937961 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="284a08f7-5dcd-48f1-823e-2c4b4e8934da" containerName="gather" Dec 02 03:14:28 crc kubenswrapper[4884]: I1202 03:14:28.941397 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:28 crc kubenswrapper[4884]: I1202 03:14:28.974067 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t88cv"] Dec 02 03:14:29 crc kubenswrapper[4884]: I1202 03:14:29.050710 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13b4bc27-0472-4566-94fe-067318344cbc-utilities\") pod \"community-operators-t88cv\" (UID: \"13b4bc27-0472-4566-94fe-067318344cbc\") " pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:29 crc kubenswrapper[4884]: I1202 03:14:29.050813 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqwh6\" (UniqueName: \"kubernetes.io/projected/13b4bc27-0472-4566-94fe-067318344cbc-kube-api-access-dqwh6\") pod \"community-operators-t88cv\" (UID: \"13b4bc27-0472-4566-94fe-067318344cbc\") " pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:29 crc kubenswrapper[4884]: I1202 03:14:29.050859 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13b4bc27-0472-4566-94fe-067318344cbc-catalog-content\") pod \"community-operators-t88cv\" (UID: \"13b4bc27-0472-4566-94fe-067318344cbc\") " pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:29 crc kubenswrapper[4884]: I1202 03:14:29.152781 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13b4bc27-0472-4566-94fe-067318344cbc-utilities\") pod \"community-operators-t88cv\" (UID: \"13b4bc27-0472-4566-94fe-067318344cbc\") " pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:29 crc kubenswrapper[4884]: I1202 03:14:29.152844 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqwh6\" (UniqueName: \"kubernetes.io/projected/13b4bc27-0472-4566-94fe-067318344cbc-kube-api-access-dqwh6\") pod \"community-operators-t88cv\" (UID: \"13b4bc27-0472-4566-94fe-067318344cbc\") " pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:29 crc kubenswrapper[4884]: I1202 03:14:29.152875 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13b4bc27-0472-4566-94fe-067318344cbc-catalog-content\") pod \"community-operators-t88cv\" (UID: \"13b4bc27-0472-4566-94fe-067318344cbc\") " pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:29 crc kubenswrapper[4884]: I1202 03:14:29.153222 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13b4bc27-0472-4566-94fe-067318344cbc-utilities\") pod \"community-operators-t88cv\" (UID: \"13b4bc27-0472-4566-94fe-067318344cbc\") " pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:29 crc kubenswrapper[4884]: I1202 03:14:29.153489 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13b4bc27-0472-4566-94fe-067318344cbc-catalog-content\") pod \"community-operators-t88cv\" (UID: \"13b4bc27-0472-4566-94fe-067318344cbc\") " pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:29 crc kubenswrapper[4884]: I1202 03:14:29.180219 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqwh6\" (UniqueName: \"kubernetes.io/projected/13b4bc27-0472-4566-94fe-067318344cbc-kube-api-access-dqwh6\") pod \"community-operators-t88cv\" (UID: \"13b4bc27-0472-4566-94fe-067318344cbc\") " pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:29 crc kubenswrapper[4884]: I1202 03:14:29.292416 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:29 crc kubenswrapper[4884]: I1202 03:14:29.849711 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t88cv"] Dec 02 03:14:30 crc kubenswrapper[4884]: I1202 03:14:30.616478 4884 generic.go:334] "Generic (PLEG): container finished" podID="13b4bc27-0472-4566-94fe-067318344cbc" containerID="fe3afd7bd932099c9ad414db3649a262d3844faf66f88566e120705bf400892c" exitCode=0 Dec 02 03:14:30 crc kubenswrapper[4884]: I1202 03:14:30.616802 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t88cv" event={"ID":"13b4bc27-0472-4566-94fe-067318344cbc","Type":"ContainerDied","Data":"fe3afd7bd932099c9ad414db3649a262d3844faf66f88566e120705bf400892c"} Dec 02 03:14:30 crc kubenswrapper[4884]: I1202 03:14:30.616830 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t88cv" event={"ID":"13b4bc27-0472-4566-94fe-067318344cbc","Type":"ContainerStarted","Data":"925159d3b229c9fa2a0eab88b64be380b6e8604a4e129554674a12cfeef8f99e"} Dec 02 03:14:30 crc kubenswrapper[4884]: I1202 03:14:30.618649 4884 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 03:14:32 crc kubenswrapper[4884]: I1202 03:14:32.648299 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t88cv" event={"ID":"13b4bc27-0472-4566-94fe-067318344cbc","Type":"ContainerStarted","Data":"91d9191f2ebfe296791520d247fe7bd982300cd48567171ca3f3a4d21c9aaf08"} Dec 02 03:14:33 crc kubenswrapper[4884]: I1202 03:14:33.662679 4884 generic.go:334] "Generic (PLEG): container finished" podID="13b4bc27-0472-4566-94fe-067318344cbc" containerID="91d9191f2ebfe296791520d247fe7bd982300cd48567171ca3f3a4d21c9aaf08" exitCode=0 Dec 02 03:14:33 crc kubenswrapper[4884]: I1202 03:14:33.662972 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t88cv" event={"ID":"13b4bc27-0472-4566-94fe-067318344cbc","Type":"ContainerDied","Data":"91d9191f2ebfe296791520d247fe7bd982300cd48567171ca3f3a4d21c9aaf08"} Dec 02 03:14:34 crc kubenswrapper[4884]: I1202 03:14:34.674523 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t88cv" event={"ID":"13b4bc27-0472-4566-94fe-067318344cbc","Type":"ContainerStarted","Data":"c9a9b9e238dd80db8f40d5afa77a959db023704315acf405b2c8c87a451854af"} Dec 02 03:14:34 crc kubenswrapper[4884]: I1202 03:14:34.702926 4884 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t88cv" podStartSLOduration=3.044925071 podStartE2EDuration="6.702906406s" podCreationTimestamp="2025-12-02 03:14:28 +0000 UTC" firstStartedPulling="2025-12-02 03:14:30.618415845 +0000 UTC m=+5887.294252729" lastFinishedPulling="2025-12-02 03:14:34.27639717 +0000 UTC m=+5890.952234064" observedRunningTime="2025-12-02 03:14:34.694022779 +0000 UTC m=+5891.369859663" watchObservedRunningTime="2025-12-02 03:14:34.702906406 +0000 UTC m=+5891.378743290" Dec 02 03:14:39 crc kubenswrapper[4884]: I1202 03:14:39.292785 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:39 crc kubenswrapper[4884]: I1202 03:14:39.293465 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:39 crc kubenswrapper[4884]: I1202 03:14:39.381396 4884 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:39 crc kubenswrapper[4884]: I1202 03:14:39.798444 4884 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:39 crc kubenswrapper[4884]: I1202 03:14:39.875053 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t88cv"] Dec 02 03:14:41 crc kubenswrapper[4884]: I1202 03:14:41.760104 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t88cv" podUID="13b4bc27-0472-4566-94fe-067318344cbc" containerName="registry-server" containerID="cri-o://c9a9b9e238dd80db8f40d5afa77a959db023704315acf405b2c8c87a451854af" gracePeriod=2 Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.304156 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.479083 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13b4bc27-0472-4566-94fe-067318344cbc-utilities\") pod \"13b4bc27-0472-4566-94fe-067318344cbc\" (UID: \"13b4bc27-0472-4566-94fe-067318344cbc\") " Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.479294 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqwh6\" (UniqueName: \"kubernetes.io/projected/13b4bc27-0472-4566-94fe-067318344cbc-kube-api-access-dqwh6\") pod \"13b4bc27-0472-4566-94fe-067318344cbc\" (UID: \"13b4bc27-0472-4566-94fe-067318344cbc\") " Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.479389 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13b4bc27-0472-4566-94fe-067318344cbc-catalog-content\") pod \"13b4bc27-0472-4566-94fe-067318344cbc\" (UID: \"13b4bc27-0472-4566-94fe-067318344cbc\") " Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.480953 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13b4bc27-0472-4566-94fe-067318344cbc-utilities" (OuterVolumeSpecName: "utilities") pod "13b4bc27-0472-4566-94fe-067318344cbc" (UID: "13b4bc27-0472-4566-94fe-067318344cbc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.486258 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13b4bc27-0472-4566-94fe-067318344cbc-kube-api-access-dqwh6" (OuterVolumeSpecName: "kube-api-access-dqwh6") pod "13b4bc27-0472-4566-94fe-067318344cbc" (UID: "13b4bc27-0472-4566-94fe-067318344cbc"). InnerVolumeSpecName "kube-api-access-dqwh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.565133 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13b4bc27-0472-4566-94fe-067318344cbc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "13b4bc27-0472-4566-94fe-067318344cbc" (UID: "13b4bc27-0472-4566-94fe-067318344cbc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.582594 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqwh6\" (UniqueName: \"kubernetes.io/projected/13b4bc27-0472-4566-94fe-067318344cbc-kube-api-access-dqwh6\") on node \"crc\" DevicePath \"\"" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.582634 4884 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13b4bc27-0472-4566-94fe-067318344cbc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.582646 4884 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13b4bc27-0472-4566-94fe-067318344cbc-utilities\") on node \"crc\" DevicePath \"\"" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.772847 4884 generic.go:334] "Generic (PLEG): container finished" podID="13b4bc27-0472-4566-94fe-067318344cbc" containerID="c9a9b9e238dd80db8f40d5afa77a959db023704315acf405b2c8c87a451854af" exitCode=0 Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.772890 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t88cv" event={"ID":"13b4bc27-0472-4566-94fe-067318344cbc","Type":"ContainerDied","Data":"c9a9b9e238dd80db8f40d5afa77a959db023704315acf405b2c8c87a451854af"} Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.772922 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t88cv" event={"ID":"13b4bc27-0472-4566-94fe-067318344cbc","Type":"ContainerDied","Data":"925159d3b229c9fa2a0eab88b64be380b6e8604a4e129554674a12cfeef8f99e"} Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.772939 4884 scope.go:117] "RemoveContainer" containerID="c9a9b9e238dd80db8f40d5afa77a959db023704315acf405b2c8c87a451854af" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.772969 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t88cv" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.807155 4884 scope.go:117] "RemoveContainer" containerID="91d9191f2ebfe296791520d247fe7bd982300cd48567171ca3f3a4d21c9aaf08" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.813962 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t88cv"] Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.825761 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t88cv"] Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.834443 4884 scope.go:117] "RemoveContainer" containerID="fe3afd7bd932099c9ad414db3649a262d3844faf66f88566e120705bf400892c" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.884320 4884 scope.go:117] "RemoveContainer" containerID="c9a9b9e238dd80db8f40d5afa77a959db023704315acf405b2c8c87a451854af" Dec 02 03:14:42 crc kubenswrapper[4884]: E1202 03:14:42.896675 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9a9b9e238dd80db8f40d5afa77a959db023704315acf405b2c8c87a451854af\": container with ID starting with c9a9b9e238dd80db8f40d5afa77a959db023704315acf405b2c8c87a451854af not found: ID does not exist" containerID="c9a9b9e238dd80db8f40d5afa77a959db023704315acf405b2c8c87a451854af" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.896943 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9a9b9e238dd80db8f40d5afa77a959db023704315acf405b2c8c87a451854af"} err="failed to get container status \"c9a9b9e238dd80db8f40d5afa77a959db023704315acf405b2c8c87a451854af\": rpc error: code = NotFound desc = could not find container \"c9a9b9e238dd80db8f40d5afa77a959db023704315acf405b2c8c87a451854af\": container with ID starting with c9a9b9e238dd80db8f40d5afa77a959db023704315acf405b2c8c87a451854af not found: ID does not exist" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.896974 4884 scope.go:117] "RemoveContainer" containerID="91d9191f2ebfe296791520d247fe7bd982300cd48567171ca3f3a4d21c9aaf08" Dec 02 03:14:42 crc kubenswrapper[4884]: E1202 03:14:42.897893 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91d9191f2ebfe296791520d247fe7bd982300cd48567171ca3f3a4d21c9aaf08\": container with ID starting with 91d9191f2ebfe296791520d247fe7bd982300cd48567171ca3f3a4d21c9aaf08 not found: ID does not exist" containerID="91d9191f2ebfe296791520d247fe7bd982300cd48567171ca3f3a4d21c9aaf08" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.897944 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91d9191f2ebfe296791520d247fe7bd982300cd48567171ca3f3a4d21c9aaf08"} err="failed to get container status \"91d9191f2ebfe296791520d247fe7bd982300cd48567171ca3f3a4d21c9aaf08\": rpc error: code = NotFound desc = could not find container \"91d9191f2ebfe296791520d247fe7bd982300cd48567171ca3f3a4d21c9aaf08\": container with ID starting with 91d9191f2ebfe296791520d247fe7bd982300cd48567171ca3f3a4d21c9aaf08 not found: ID does not exist" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.897979 4884 scope.go:117] "RemoveContainer" containerID="fe3afd7bd932099c9ad414db3649a262d3844faf66f88566e120705bf400892c" Dec 02 03:14:42 crc kubenswrapper[4884]: E1202 03:14:42.898408 4884 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe3afd7bd932099c9ad414db3649a262d3844faf66f88566e120705bf400892c\": container with ID starting with fe3afd7bd932099c9ad414db3649a262d3844faf66f88566e120705bf400892c not found: ID does not exist" containerID="fe3afd7bd932099c9ad414db3649a262d3844faf66f88566e120705bf400892c" Dec 02 03:14:42 crc kubenswrapper[4884]: I1202 03:14:42.898435 4884 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe3afd7bd932099c9ad414db3649a262d3844faf66f88566e120705bf400892c"} err="failed to get container status \"fe3afd7bd932099c9ad414db3649a262d3844faf66f88566e120705bf400892c\": rpc error: code = NotFound desc = could not find container \"fe3afd7bd932099c9ad414db3649a262d3844faf66f88566e120705bf400892c\": container with ID starting with fe3afd7bd932099c9ad414db3649a262d3844faf66f88566e120705bf400892c not found: ID does not exist" Dec 02 03:14:43 crc kubenswrapper[4884]: I1202 03:14:43.633376 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13b4bc27-0472-4566-94fe-067318344cbc" path="/var/lib/kubelet/pods/13b4bc27-0472-4566-94fe-067318344cbc/volumes" Dec 02 03:14:48 crc kubenswrapper[4884]: I1202 03:14:48.245301 4884 scope.go:117] "RemoveContainer" containerID="1032ec5163c2492db9cacf75bd0ea2eec65d5903dfa4b08fdb4d2ceb746c466f" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.154350 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh"] Dec 02 03:15:00 crc kubenswrapper[4884]: E1202 03:15:00.155396 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13b4bc27-0472-4566-94fe-067318344cbc" containerName="registry-server" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.155413 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="13b4bc27-0472-4566-94fe-067318344cbc" containerName="registry-server" Dec 02 03:15:00 crc kubenswrapper[4884]: E1202 03:15:00.155438 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13b4bc27-0472-4566-94fe-067318344cbc" containerName="extract-content" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.155452 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="13b4bc27-0472-4566-94fe-067318344cbc" containerName="extract-content" Dec 02 03:15:00 crc kubenswrapper[4884]: E1202 03:15:00.155489 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13b4bc27-0472-4566-94fe-067318344cbc" containerName="extract-utilities" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.155501 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="13b4bc27-0472-4566-94fe-067318344cbc" containerName="extract-utilities" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.155789 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="13b4bc27-0472-4566-94fe-067318344cbc" containerName="registry-server" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.156717 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.159732 4884 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.160370 4884 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.168768 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh"] Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.262831 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7117324b-6bf3-4faa-a3a0-77cff57bd01a-secret-volume\") pod \"collect-profiles-29410755-zkxlh\" (UID: \"7117324b-6bf3-4faa-a3a0-77cff57bd01a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.263269 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sdxs\" (UniqueName: \"kubernetes.io/projected/7117324b-6bf3-4faa-a3a0-77cff57bd01a-kube-api-access-7sdxs\") pod \"collect-profiles-29410755-zkxlh\" (UID: \"7117324b-6bf3-4faa-a3a0-77cff57bd01a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.263392 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7117324b-6bf3-4faa-a3a0-77cff57bd01a-config-volume\") pod \"collect-profiles-29410755-zkxlh\" (UID: \"7117324b-6bf3-4faa-a3a0-77cff57bd01a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.364910 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7117324b-6bf3-4faa-a3a0-77cff57bd01a-secret-volume\") pod \"collect-profiles-29410755-zkxlh\" (UID: \"7117324b-6bf3-4faa-a3a0-77cff57bd01a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.365070 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sdxs\" (UniqueName: \"kubernetes.io/projected/7117324b-6bf3-4faa-a3a0-77cff57bd01a-kube-api-access-7sdxs\") pod \"collect-profiles-29410755-zkxlh\" (UID: \"7117324b-6bf3-4faa-a3a0-77cff57bd01a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.365110 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7117324b-6bf3-4faa-a3a0-77cff57bd01a-config-volume\") pod \"collect-profiles-29410755-zkxlh\" (UID: \"7117324b-6bf3-4faa-a3a0-77cff57bd01a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.366409 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7117324b-6bf3-4faa-a3a0-77cff57bd01a-config-volume\") pod \"collect-profiles-29410755-zkxlh\" (UID: \"7117324b-6bf3-4faa-a3a0-77cff57bd01a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.372347 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7117324b-6bf3-4faa-a3a0-77cff57bd01a-secret-volume\") pod \"collect-profiles-29410755-zkxlh\" (UID: \"7117324b-6bf3-4faa-a3a0-77cff57bd01a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.392534 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sdxs\" (UniqueName: \"kubernetes.io/projected/7117324b-6bf3-4faa-a3a0-77cff57bd01a-kube-api-access-7sdxs\") pod \"collect-profiles-29410755-zkxlh\" (UID: \"7117324b-6bf3-4faa-a3a0-77cff57bd01a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" Dec 02 03:15:00 crc kubenswrapper[4884]: I1202 03:15:00.519970 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" Dec 02 03:15:01 crc kubenswrapper[4884]: I1202 03:15:01.082089 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh"] Dec 02 03:15:02 crc kubenswrapper[4884]: I1202 03:15:02.014660 4884 generic.go:334] "Generic (PLEG): container finished" podID="7117324b-6bf3-4faa-a3a0-77cff57bd01a" containerID="535d4f1f2f691410774bf813c1fe109f5661569ba0164350cb2bd419c931e5c7" exitCode=0 Dec 02 03:15:02 crc kubenswrapper[4884]: I1202 03:15:02.014774 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" event={"ID":"7117324b-6bf3-4faa-a3a0-77cff57bd01a","Type":"ContainerDied","Data":"535d4f1f2f691410774bf813c1fe109f5661569ba0164350cb2bd419c931e5c7"} Dec 02 03:15:02 crc kubenswrapper[4884]: I1202 03:15:02.015097 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" event={"ID":"7117324b-6bf3-4faa-a3a0-77cff57bd01a","Type":"ContainerStarted","Data":"24ba3284813940f583024e7bb5e6a74f046d41b05202701233db19a1fdafa028"} Dec 02 03:15:03 crc kubenswrapper[4884]: I1202 03:15:03.451350 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" Dec 02 03:15:03 crc kubenswrapper[4884]: I1202 03:15:03.531825 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7sdxs\" (UniqueName: \"kubernetes.io/projected/7117324b-6bf3-4faa-a3a0-77cff57bd01a-kube-api-access-7sdxs\") pod \"7117324b-6bf3-4faa-a3a0-77cff57bd01a\" (UID: \"7117324b-6bf3-4faa-a3a0-77cff57bd01a\") " Dec 02 03:15:03 crc kubenswrapper[4884]: I1202 03:15:03.532568 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7117324b-6bf3-4faa-a3a0-77cff57bd01a-secret-volume\") pod \"7117324b-6bf3-4faa-a3a0-77cff57bd01a\" (UID: \"7117324b-6bf3-4faa-a3a0-77cff57bd01a\") " Dec 02 03:15:03 crc kubenswrapper[4884]: I1202 03:15:03.532797 4884 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7117324b-6bf3-4faa-a3a0-77cff57bd01a-config-volume\") pod \"7117324b-6bf3-4faa-a3a0-77cff57bd01a\" (UID: \"7117324b-6bf3-4faa-a3a0-77cff57bd01a\") " Dec 02 03:15:03 crc kubenswrapper[4884]: I1202 03:15:03.533386 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7117324b-6bf3-4faa-a3a0-77cff57bd01a-config-volume" (OuterVolumeSpecName: "config-volume") pod "7117324b-6bf3-4faa-a3a0-77cff57bd01a" (UID: "7117324b-6bf3-4faa-a3a0-77cff57bd01a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 02 03:15:03 crc kubenswrapper[4884]: I1202 03:15:03.533706 4884 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7117324b-6bf3-4faa-a3a0-77cff57bd01a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 02 03:15:03 crc kubenswrapper[4884]: I1202 03:15:03.537304 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7117324b-6bf3-4faa-a3a0-77cff57bd01a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7117324b-6bf3-4faa-a3a0-77cff57bd01a" (UID: "7117324b-6bf3-4faa-a3a0-77cff57bd01a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 02 03:15:03 crc kubenswrapper[4884]: I1202 03:15:03.537444 4884 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7117324b-6bf3-4faa-a3a0-77cff57bd01a-kube-api-access-7sdxs" (OuterVolumeSpecName: "kube-api-access-7sdxs") pod "7117324b-6bf3-4faa-a3a0-77cff57bd01a" (UID: "7117324b-6bf3-4faa-a3a0-77cff57bd01a"). InnerVolumeSpecName "kube-api-access-7sdxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 02 03:15:03 crc kubenswrapper[4884]: I1202 03:15:03.636476 4884 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7sdxs\" (UniqueName: \"kubernetes.io/projected/7117324b-6bf3-4faa-a3a0-77cff57bd01a-kube-api-access-7sdxs\") on node \"crc\" DevicePath \"\"" Dec 02 03:15:03 crc kubenswrapper[4884]: I1202 03:15:03.636601 4884 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7117324b-6bf3-4faa-a3a0-77cff57bd01a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 02 03:15:04 crc kubenswrapper[4884]: I1202 03:15:04.042018 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" event={"ID":"7117324b-6bf3-4faa-a3a0-77cff57bd01a","Type":"ContainerDied","Data":"24ba3284813940f583024e7bb5e6a74f046d41b05202701233db19a1fdafa028"} Dec 02 03:15:04 crc kubenswrapper[4884]: I1202 03:15:04.042349 4884 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24ba3284813940f583024e7bb5e6a74f046d41b05202701233db19a1fdafa028" Dec 02 03:15:04 crc kubenswrapper[4884]: I1202 03:15:04.042125 4884 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29410755-zkxlh" Dec 02 03:15:04 crc kubenswrapper[4884]: I1202 03:15:04.586433 4884 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx"] Dec 02 03:15:04 crc kubenswrapper[4884]: I1202 03:15:04.621523 4884 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29410710-7lppx"] Dec 02 03:15:05 crc kubenswrapper[4884]: I1202 03:15:05.628086 4884 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5" path="/var/lib/kubelet/pods/ddf3d6bf-05f6-4097-9b4f-c7d5df4cffd5/volumes" Dec 02 03:15:16 crc kubenswrapper[4884]: I1202 03:15:16.971868 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 03:15:16 crc kubenswrapper[4884]: I1202 03:15:16.972488 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 03:15:46 crc kubenswrapper[4884]: I1202 03:15:46.972179 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 03:15:46 crc kubenswrapper[4884]: I1202 03:15:46.972892 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 03:15:48 crc kubenswrapper[4884]: I1202 03:15:48.371168 4884 scope.go:117] "RemoveContainer" containerID="4f9e7fa26d396a1e88176af67d53c99f72b00a3e13625da16517f8f000adf4ce" Dec 02 03:16:16 crc kubenswrapper[4884]: I1202 03:16:16.972276 4884 patch_prober.go:28] interesting pod/machine-config-daemon-jc7vx container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 02 03:16:16 crc kubenswrapper[4884]: I1202 03:16:16.972795 4884 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 02 03:16:16 crc kubenswrapper[4884]: I1202 03:16:16.972850 4884 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" Dec 02 03:16:16 crc kubenswrapper[4884]: I1202 03:16:16.973773 4884 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc"} pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 02 03:16:16 crc kubenswrapper[4884]: I1202 03:16:16.973838 4884 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerName="machine-config-daemon" containerID="cri-o://056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" gracePeriod=600 Dec 02 03:16:17 crc kubenswrapper[4884]: E1202 03:16:17.104725 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:16:18 crc kubenswrapper[4884]: I1202 03:16:18.023509 4884 generic.go:334] "Generic (PLEG): container finished" podID="9385a5c9-a156-460a-8ca1-9b342ab835c9" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" exitCode=0 Dec 02 03:16:18 crc kubenswrapper[4884]: I1202 03:16:18.023590 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" event={"ID":"9385a5c9-a156-460a-8ca1-9b342ab835c9","Type":"ContainerDied","Data":"056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc"} Dec 02 03:16:18 crc kubenswrapper[4884]: I1202 03:16:18.024103 4884 scope.go:117] "RemoveContainer" containerID="783609d4dac9b374973e24fed69e04030080c197a1b2e0ea5d35a59a53de30f9" Dec 02 03:16:18 crc kubenswrapper[4884]: I1202 03:16:18.025359 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:16:18 crc kubenswrapper[4884]: E1202 03:16:18.026547 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:16:30 crc kubenswrapper[4884]: I1202 03:16:30.614431 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:16:30 crc kubenswrapper[4884]: E1202 03:16:30.615425 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:16:45 crc kubenswrapper[4884]: I1202 03:16:45.614782 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:16:45 crc kubenswrapper[4884]: E1202 03:16:45.616028 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:16:59 crc kubenswrapper[4884]: I1202 03:16:59.615520 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:16:59 crc kubenswrapper[4884]: E1202 03:16:59.616649 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:17:10 crc kubenswrapper[4884]: I1202 03:17:10.614101 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:17:10 crc kubenswrapper[4884]: E1202 03:17:10.616864 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:17:21 crc kubenswrapper[4884]: I1202 03:17:21.614375 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:17:21 crc kubenswrapper[4884]: E1202 03:17:21.615130 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:17:32 crc kubenswrapper[4884]: I1202 03:17:32.614481 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:17:32 crc kubenswrapper[4884]: E1202 03:17:32.615326 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:17:47 crc kubenswrapper[4884]: I1202 03:17:47.614924 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:17:47 crc kubenswrapper[4884]: E1202 03:17:47.618083 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:18:02 crc kubenswrapper[4884]: I1202 03:18:02.614798 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:18:02 crc kubenswrapper[4884]: E1202 03:18:02.615572 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:18:15 crc kubenswrapper[4884]: I1202 03:18:15.614930 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:18:15 crc kubenswrapper[4884]: E1202 03:18:15.615979 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:18:30 crc kubenswrapper[4884]: I1202 03:18:30.614366 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:18:30 crc kubenswrapper[4884]: E1202 03:18:30.615135 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:18:43 crc kubenswrapper[4884]: I1202 03:18:43.627395 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:18:43 crc kubenswrapper[4884]: E1202 03:18:43.628469 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:18:56 crc kubenswrapper[4884]: I1202 03:18:56.614292 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:18:56 crc kubenswrapper[4884]: E1202 03:18:56.615371 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:19:11 crc kubenswrapper[4884]: I1202 03:19:11.616217 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:19:11 crc kubenswrapper[4884]: E1202 03:19:11.617828 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:19:26 crc kubenswrapper[4884]: I1202 03:19:26.615255 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:19:26 crc kubenswrapper[4884]: E1202 03:19:26.616445 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:19:37 crc kubenswrapper[4884]: I1202 03:19:37.614434 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:19:37 crc kubenswrapper[4884]: E1202 03:19:37.615293 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:19:48 crc kubenswrapper[4884]: I1202 03:19:48.615740 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:19:48 crc kubenswrapper[4884]: E1202 03:19:48.617118 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:20:03 crc kubenswrapper[4884]: I1202 03:20:03.633113 4884 scope.go:117] "RemoveContainer" containerID="056900d90268aef2e3ef2018a072fac257b8483e7064e8d54e4e6006581cefbc" Dec 02 03:20:03 crc kubenswrapper[4884]: E1202 03:20:03.634089 4884 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-jc7vx_openshift-machine-config-operator(9385a5c9-a156-460a-8ca1-9b342ab835c9)\"" pod="openshift-machine-config-operator/machine-config-daemon-jc7vx" podUID="9385a5c9-a156-460a-8ca1-9b342ab835c9" Dec 02 03:20:04 crc kubenswrapper[4884]: I1202 03:20:04.787381 4884 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h7px7"] Dec 02 03:20:04 crc kubenswrapper[4884]: E1202 03:20:04.791646 4884 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7117324b-6bf3-4faa-a3a0-77cff57bd01a" containerName="collect-profiles" Dec 02 03:20:04 crc kubenswrapper[4884]: I1202 03:20:04.791674 4884 state_mem.go:107] "Deleted CPUSet assignment" podUID="7117324b-6bf3-4faa-a3a0-77cff57bd01a" containerName="collect-profiles" Dec 02 03:20:04 crc kubenswrapper[4884]: I1202 03:20:04.792126 4884 memory_manager.go:354] "RemoveStaleState removing state" podUID="7117324b-6bf3-4faa-a3a0-77cff57bd01a" containerName="collect-profiles" Dec 02 03:20:04 crc kubenswrapper[4884]: I1202 03:20:04.798619 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h7px7" Dec 02 03:20:04 crc kubenswrapper[4884]: I1202 03:20:04.877165 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h7px7"] Dec 02 03:20:04 crc kubenswrapper[4884]: I1202 03:20:04.977766 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlkgg\" (UniqueName: \"kubernetes.io/projected/ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab-kube-api-access-nlkgg\") pod \"redhat-marketplace-h7px7\" (UID: \"ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab\") " pod="openshift-marketplace/redhat-marketplace-h7px7" Dec 02 03:20:04 crc kubenswrapper[4884]: I1202 03:20:04.977910 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab-utilities\") pod \"redhat-marketplace-h7px7\" (UID: \"ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab\") " pod="openshift-marketplace/redhat-marketplace-h7px7" Dec 02 03:20:04 crc kubenswrapper[4884]: I1202 03:20:04.977964 4884 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab-catalog-content\") pod \"redhat-marketplace-h7px7\" (UID: \"ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab\") " pod="openshift-marketplace/redhat-marketplace-h7px7" Dec 02 03:20:05 crc kubenswrapper[4884]: I1202 03:20:05.079305 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab-utilities\") pod \"redhat-marketplace-h7px7\" (UID: \"ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab\") " pod="openshift-marketplace/redhat-marketplace-h7px7" Dec 02 03:20:05 crc kubenswrapper[4884]: I1202 03:20:05.079391 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab-catalog-content\") pod \"redhat-marketplace-h7px7\" (UID: \"ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab\") " pod="openshift-marketplace/redhat-marketplace-h7px7" Dec 02 03:20:05 crc kubenswrapper[4884]: I1202 03:20:05.079451 4884 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlkgg\" (UniqueName: \"kubernetes.io/projected/ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab-kube-api-access-nlkgg\") pod \"redhat-marketplace-h7px7\" (UID: \"ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab\") " pod="openshift-marketplace/redhat-marketplace-h7px7" Dec 02 03:20:05 crc kubenswrapper[4884]: I1202 03:20:05.079734 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab-utilities\") pod \"redhat-marketplace-h7px7\" (UID: \"ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab\") " pod="openshift-marketplace/redhat-marketplace-h7px7" Dec 02 03:20:05 crc kubenswrapper[4884]: I1202 03:20:05.080081 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab-catalog-content\") pod \"redhat-marketplace-h7px7\" (UID: \"ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab\") " pod="openshift-marketplace/redhat-marketplace-h7px7" Dec 02 03:20:05 crc kubenswrapper[4884]: I1202 03:20:05.111397 4884 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlkgg\" (UniqueName: \"kubernetes.io/projected/ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab-kube-api-access-nlkgg\") pod \"redhat-marketplace-h7px7\" (UID: \"ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab\") " pod="openshift-marketplace/redhat-marketplace-h7px7" Dec 02 03:20:05 crc kubenswrapper[4884]: I1202 03:20:05.143983 4884 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h7px7" Dec 02 03:20:05 crc kubenswrapper[4884]: I1202 03:20:05.664488 4884 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h7px7"] Dec 02 03:20:05 crc kubenswrapper[4884]: W1202 03:20:05.680261 4884 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce7a9cd0_f67b_4df7_baaf_a2ea7e8046ab.slice/crio-064bb25aeb7e8a27fc9a4d6f090f0c19f27997953a2625f1bebe6634e481b248 WatchSource:0}: Error finding container 064bb25aeb7e8a27fc9a4d6f090f0c19f27997953a2625f1bebe6634e481b248: Status 404 returned error can't find the container with id 064bb25aeb7e8a27fc9a4d6f090f0c19f27997953a2625f1bebe6634e481b248 Dec 02 03:20:06 crc kubenswrapper[4884]: I1202 03:20:06.243965 4884 generic.go:334] "Generic (PLEG): container finished" podID="ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab" containerID="570c5fcbfa692ac33c14afb4d83887e3315615c4905b1cf568eefb39eb227393" exitCode=0 Dec 02 03:20:06 crc kubenswrapper[4884]: I1202 03:20:06.244315 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h7px7" event={"ID":"ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab","Type":"ContainerDied","Data":"570c5fcbfa692ac33c14afb4d83887e3315615c4905b1cf568eefb39eb227393"} Dec 02 03:20:06 crc kubenswrapper[4884]: I1202 03:20:06.244390 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h7px7" event={"ID":"ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab","Type":"ContainerStarted","Data":"064bb25aeb7e8a27fc9a4d6f090f0c19f27997953a2625f1bebe6634e481b248"} Dec 02 03:20:06 crc kubenswrapper[4884]: I1202 03:20:06.247565 4884 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 02 03:20:08 crc kubenswrapper[4884]: I1202 03:20:08.272505 4884 generic.go:334] "Generic (PLEG): container finished" podID="ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab" containerID="809b8fb8f4b1d781fd294901a684963e86eac2985e85ca01972f8f66cc2d8612" exitCode=0 Dec 02 03:20:08 crc kubenswrapper[4884]: I1202 03:20:08.272832 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h7px7" event={"ID":"ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab","Type":"ContainerDied","Data":"809b8fb8f4b1d781fd294901a684963e86eac2985e85ca01972f8f66cc2d8612"} Dec 02 03:20:09 crc kubenswrapper[4884]: I1202 03:20:09.287501 4884 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h7px7" event={"ID":"ce7a9cd0-f67b-4df7-baaf-a2ea7e8046ab","Type":"ContainerStarted","Data":"8063e74086b6837598994a69ef153707cf6d7f6fc09f019cc9280021076f5963"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515113455361024451 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015113455362017367 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015113440661016506 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015113440661015456 5ustar corecore